Introduction to Bayesian methods in inverse problems
|
|
- Quentin Walters
- 6 years ago
- Views:
Transcription
1 Introduction to Bayesian methods in inverse problems Ville Kolehmainen 1 1 Department of Applied Physics, University of Eastern Finland, Kuopio, Finland March Manchester, UK.
2 Contents Introduction Bayesian inversion Computation of integration based estimates Numerical examples
3 What are inverse problems? Consider measurement problems; estimate the quantity of interest x R n from (noisy) measurement of A(x) R d, where A is known mapping. Inverse problems are characterized as those measurement problems which are ill-posed: 1. The problem is non-unique (e.g., less measurements than unknowns) 2. Solution is unstable w.r.t measurement noise and modelling errors (e.g., model reduction, inaccurately known nuisance parameters in the model A(x)).
4 An introductory example: Consider 2D deconvolution (image deblurring); Given noisy and blurred image m = Ax + e, m R n the objective is to reconstruct the original image x R n. Forward model x Ax implements discrete convolution (here the convolution kernel is Gaussian blurring kernel with std of 6 pixels) Left; Original image x. Right; Observed image m = Ax + e.
5 Example (cont.) Matrix A has trivial null-space null(a) = {0} solution is unique (i.e. (A T A) 1 ). However, the problem is unstable (A has "unclear" null-space, i.e., Aw = k σ k w, v k u k 0 for certain w = 1) Figure shows the least squares (LS) solution x LS = arg min x m Ax 2 x LS = (A T A) 1 A T m Left; true image x. Right; LS solution x LS
6 Solutions; Regularization. The ill posed problem is replaced by a well posed approximation. Solution hopefully close to the true solution. Typically modifications of the associated LS problem m Ax 2. Examples of methods; Tikhonov regularization, truncated iterations. Consider the LS problem x LS = arg min{ m Ax 2 x 2 } (AT A) x }{{} LS = A T m Uniqueness; B 1 if null(a) = {0}. Stability of the solution? B
7 Regularization Example (cont.); In Tikhonov regularization, the LS problem is replaced with x TIK = arg min{ m Ax 2 x 2 +α x 2 2 } (AT A + αi) x }{{} TIK = A T m Uniqueness; α > 0 null( B) = {0} B 1. Stability of the solution guaranteed by choosing α large enough. Regularization poses (implicit) prior about the solution. These assumptions are sometimes well hidden. B Left; true image x. Right; Tikhonov solution x TIK.
8 Solutions; Statistical (Bayesian) inversion. The inverse problem is recast as a problem of Bayesian inference. The key idea is to extract estimates and assess uncertainty about the unknowns based on Measured data Model of measurement process Model of a priori information Ill-posedness removed by explicit use of prior models! Systematic handling of model uncertainties and reduction ( approximation error theory) Left; true image x. Right; Bayesian estimate x CM.
9 Bayesian inversion. All variables in the model are considered as random variables. The randomness reflects our uncertainty of their actual values. The degree of uncertainty is coded in the probability distribution models of these variables. The complete model of the inverse problem is the posterior probability distribution π(x m) = π pr(x)π(m x), m = m observed (1) π(m) where π(m x) is the likelihood density; model of the measurement process. π pr (x) is the prior density; model for a priori information. π(m) is a normalization constant
10 The posterior density model is a function on n dimensional space; π(x m) : R n R + where n is usually large. To obtain "practical solution", the posterior model is summarized by estimates that answer to questions such as; What is the most probable value of x over π(x m)? What is the mean of x over π(x m)? In what interval are the values of x with 90% (posterior) probability?
11 Typical point estimates Maximum a posteriori (MAP) estimate: π(x MAP m) = arg max π(x m). x Conditional mean (CM) estimate: x CM = xπ(x m)dx. R n Covariance: Γ x m = (x x CM )(x x CM ) T π(x m) dx R n
12 Confidence intervals; Given 0 < τ < 100, compute a k and b k s.t. ak π k (x k ) dx k = b k π k (x k ) dx k = 100 τ 200 where π k (x k ) is the marginal density π k (x k ) = π(x m) dx 1 dx k 1 dx k+1 dx n. R n 1 The interval I k (τ) = [a k b k ] contains τ % of the mass of the marginal density.
13 Computation of the integration based estimates Many estimators are of the form f (x) = R n f (x)π(x m)dx Examples: f (x) = x x CM f (x) = (x x CM )(x x CM ) T Γ x m etc... Analytical evaluation in most cases impossible Traditional numerical quadratures not applicable when n is large (number of points needed unreasonably large, support of π(x m) may not be well known) Monte Carlo integration.
14 Monte Carlo integration Monte Carlo integration 1. Draw an ensemble {x (k), k = 1, 2,..., N} of i.i.d samples from π(x). 2. Estimate R n f (x)π(x)dx 1 N N f (x (k) ) k=1 Convergence (law of large numbers) 1 N lim N f (x (k) ) f (x)π(x)dx N R n k=1 (a.c) Variance of the estimator f = 1 N N k=1 f (x (k) ) reduces 1 N
15 Simple example of Monte Carlo integration Let x R 2 and { 1 π(x) = 4, x G, G = [ 1, 1] [ 1, 1] 0, x / G. The task is to compute integral g(x) = R 2 χ D π(x)dx where D is the unit disk on R 2. Using N = 5000 samples, we get estimate g(x) = (true value π/4 = )
16 MCMC Direct sampling of the posterior usually not possible Markov chain Monte Carlo (MCMC). MCMC is Monte Carlo integration using Markov chains (dependent samples); 1. Draw {x (k), k = 1, 2,..., N} π(x) by simulating a Markov chain (with equilibrium distribution π(x)). 2. Estimate f (x)π(x)dx 1 N f (x (k) ) R N n Variance of the estimator reduces as τ/n where τ is the integrated autocorrelation time of the chain. Algorithms for MCMC; 1. Metropolis Hastings algorithm 2. Gibbs sampler k=1
17 Metropolis-Hastings algorihtm Generation of the ensemble {x (k), k = 1,..., N} π(x) using Metropolis Hastings algorithm; 1. Pick an initial value x (1) R n and set l = 1 2. Set x = x (l). 3. Draw a candidate sample x from proposal density x q(x, x ) and compute the acceptance factor ( α(x, x ) = min 1, π(x )q(x ), x) π(x)q(x, x. ) 4. Draw t [0, 1] from uniform probability density (t uni(0, 1)). 5. If α(x, x ) t, set x (l+1) = x, else x (l+1) = x. Increment l l When l = N stop, else repeat from step 2.
18 Normalization constant of π do not need to be known. Great flexibility in choosing the proposal density q(x, x ); almost any density would do the job (eventually). However, the choice of q(x, x ) is a crucial part of successful MCMC; it determines the efficiency (autocorrelation time τ) of the algorithm q(x, x ) should be s.t. τ as small as possible No systematic methods for choosing q(x, x ). More based on intuition & art.
19 The updating may proceed All unknowns at a time Single component x j at a time A block of unknowns at a time The order of updating (in single component and blockwise updating) may be Update elements chosen in random (random scan) Systematic order The proposal can be mixture of several densities q(x, x ) = t p i q i (x, x ), i=1 t p i = 1 i=1
20 Proposal parameteres are usually calibrated by pilot runs; aim at finding parameters giving best efficiency (minimal τ). Determining the "burn-in"; Trial runs (e.g, running several chains & checking that they are consistent). Often determined from the plot of posterior trace {π(x (k) ), k = 1,..., N} (see Figure).
21 Numerical examples In summary, Bayesian solution of an inverse problem consist of the following steps: 1. Construct the likelihood model π(m x). This step includes the development of the model x A(x) and modeling the measurement noise and modelling errors. 2. Construction of the prior model π pr (x). 3. Develop computation methods for the summary statistics. We consider the the following numerical examples: 1. 2D deconvolution (intro example continued) 2. EIT with experimental data
22 Example 1; 2D deconvolution We are given a blurred and noisy image m = Ax + e, m R n Assume that we know a priori that the true solution is binary image representing some text Left; true image x. Right; Noisy, blurred image m = Ax + e.
23 Likelihood model π(m x): Joint density π(m, x, e) = π(m x, e)π(e x)π(x) = π(m, e x)π(x) In case of m = Ax + e, we have π(m x, e) = δ(m Ax e), and π(m x) = π(m, e x) de = δ(m Ax e)π(e x) de = π e x (m Ax x)
24 In the (usual) case of mutually independent x and e, we have π e x (e x) = π e (e) and π(m x) = π e (m Ax) The model of the noise: π(e) = N (0, Γ e ), ( π(m x) exp 1 ( L e (m Ax) 2)), 2 where L T el e = Γ 1 e. Remark; we have assumed that the (numerical) model Ax is exact (i.e., no modelling errors).
25 Prior model π pr (x) The prior knowledge: x can obtain the values x j {0, 1} Case x j = 0 (black, background), case x j = 1 (white, text). The pixels with value x j = 1 are known to be clustered in "blocky structures" which have short boundaries. We model the prior knowledge with Ising prior n π pr (x) exp α δ(x i, x j ) i=1 j N i where δ is the Kronecker delta { 1, xi = x δ(x i, x j ) = j 0, x i x j
26 Posterior model π(x m): Posterior model for the deblurring problem; π(x m) exp 1 n 2 L e(m Ax) α δ(x i, x j ) i=1 j N i Posterior explored with the Metropolis Hastings algorithm.
27 Metropolis Hastings proposal The proposal is a mixture q(x, x ) = 2 i=1 ξ iq i (x, x ) of two move types. Move 1: choose update element x i R by drawing index i with uniform probability 1 n and change the value of x i. Move 2: Let N (x) denote the set of active edges in image x (edge l ij connects pixels x i and x j in the lattice; it is active if x i x j ). Pick an active update edge w.p. 1 N (x) Pick one of the two pixels related to the chosen edge w.p 1 2 and change the value of the chosen pixel.
28 Results Left image; True image x Middle; Tikhonov regularized solution x TIK = arg min x { m Ax α x 2 2 } Right image; CM estimate x CM = R n xπ(x m)dx Left; true image x, Middle; x TIK. Right; x CM.
29 Results Left image; x CM Posterior variances diag(γ x m ) Right; A sample from π(x m) Left; x CM, Middle; posterior variances diag(γ x m ). Right; A sample x from π(x m).
30 Introduction Bayesian inversion Computation of integration based estimates Numerical examples Example 2: "tank EIT" Joint work with Geoff Nicholls (Oxford) & Colin Fox (U. Otago, NZ) Data measured with the Kuopio EIT device. 16 electrodes, adjacent current patterns. Target: plastic cylinders in salt water. Measured voltages
31 Measurement model V = U(σ) + e, e N (e, Γ e ) where V denotes the measured voltages and U(σ) FEM-based forward map. e and σ modelled mutually independent. e and Γ e estimated by repeated measurements. Posterior model { π(σ V ) exp 1 } 2 (V U(σ))T Γ 1 e (V U(σ)) π pr (σ) We compute results with three different prior models π pr (σ). Sampling carried out with the Metropolis-Hasting sampling algorithm.
32 1 SMOOTHNESS PRIOR π pr (σ) π + (σ) exp Prior models π pr (σ) ( α ) n H i (σ), H i (σ) = σ i σ j 2. j N i i=1 2 MATERIAL TYPE PRIOR (FOX & NICHOLLS,1998) The possible material types {1, 2,..., C} inside the body are known but the segmentation to these materials is unknown. The material distibution is represented by an (auxliary) dicrete valued random field τ (pixel values τ j {1, 2,..., C}). The possible values of conductivity σ(τ) for different materials are known only approximately. Gaussian prior for the material conductivities. The different material types are assumed to be clustered in blocky structures (e.g., organs, etc)
33 Prior model π pr (σ) = π pr (σ τ)π pr (τ), where ( ) n n π pr (σ τ) π + (σ) exp α G i (σ) π(σ i τ i ) where π(σ i τ i ) = N (η(τ i ), ξ(τ i ) 2 ) and G i (σ) = i i=1 i=1 j {k k N i andτ k =τ i } (σ i σ j ) 2 π pr (τ) is an Ising prior: ( ) n π pr (τ) exp β T i (τ), T i (τ) = δ(τ i, τ j ), j N i (2) i=1 where δ(τ i, τ j ) is the Kronecker delta function.
34 3 CIRCLE PRIOR Domain Ω with known (constant) background conductivity σ bg is assumed to contain unknown number of circular inclusions (i.e., dimension of state not fixed) The inlcusions have known contrast Sizes and locations of inclusions unknown For the number N of the circular inclusions we write a point process prior: π pr (σ) = β N exp( βa)δ(σ A), where β = 1/A (A is the area of the domain Ω) A is set of feasible circle configurations (the circle inclusions are disjoint) and δ is the indicator function.
35 CM estimates with the different priors Smoothness prior (Top right), material type prior (Bottom left) and circle prior (Bottom right)
MCMC Sampling for Bayesian Inference using L1-type Priors
MÜNSTER MCMC Sampling for Bayesian Inference using L1-type Priors (what I do whenever the ill-posedness of EEG/MEG is just not frustrating enough!) AG Imaging Seminar Felix Lucka 26.06.2012 , MÜNSTER Sampling
More informationStatistical Estimation of the Parameters of a PDE
PIMS-MITACS 2001 1 Statistical Estimation of the Parameters of a PDE Colin Fox, Geoff Nicholls (University of Auckland) Nomenclature for image recovery Statistical model for inverse problems Traditional
More informationThe University of Auckland Applied Mathematics Bayesian Methods for Inverse Problems : why and how Colin Fox Tiangang Cui, Mike O Sullivan (Auckland),
The University of Auckland Applied Mathematics Bayesian Methods for Inverse Problems : why and how Colin Fox Tiangang Cui, Mike O Sullivan (Auckland), Geoff Nicholls (Statistics, Oxford) fox@math.auckland.ac.nz
More informationDeblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox Richard A. Norton, J.
Deblurring Jupiter (sampling in GLIP faster than regularized inversion) Colin Fox fox@physics.otago.ac.nz Richard A. Norton, J. Andrés Christen Topics... Backstory (?) Sampling in linear-gaussian hierarchical
More informationMarkov Chain Monte Carlo (MCMC)
Markov Chain Monte Carlo (MCMC Dependent Sampling Suppose we wish to sample from a density π, and we can evaluate π as a function but have no means to directly generate a sample. Rejection sampling can
More informationComputational statistics
Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated
More informationAn example of Bayesian reasoning Consider the one-dimensional deconvolution problem with various degrees of prior information.
An example of Bayesian reasoning Consider the one-dimensional deconvolution problem with various degrees of prior information. Model: where g(t) = a(t s)f(s)ds + e(t), a(t) t = (rapidly). The problem,
More informationBayesian Methods and Uncertainty Quantification for Nonlinear Inverse Problems
Bayesian Methods and Uncertainty Quantification for Nonlinear Inverse Problems John Bardsley, University of Montana Collaborators: H. Haario, J. Kaipio, M. Laine, Y. Marzouk, A. Seppänen, A. Solonen, Z.
More informationBagging During Markov Chain Monte Carlo for Smoother Predictions
Bagging During Markov Chain Monte Carlo for Smoother Predictions Herbert K. H. Lee University of California, Santa Cruz Abstract: Making good predictions from noisy data is a challenging problem. Methods
More informationA Search and Jump Algorithm for Markov Chain Monte Carlo Sampling. Christopher Jennison. Adriana Ibrahim. Seminar at University of Kuwait
A Search and Jump Algorithm for Markov Chain Monte Carlo Sampling Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj Adriana Ibrahim Institute
More informationApril 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning
for for Advanced Topics in California Institute of Technology April 20th, 2017 1 / 50 Table of Contents for 1 2 3 4 2 / 50 History of methods for Enrico Fermi used to calculate incredibly accurate predictions
More informationMarginal density. If the unknown is of the form x = (x 1, x 2 ) in which the target of investigation is x 1, a marginal posterior density
Marginal density If the unknown is of the form x = x 1, x 2 ) in which the target of investigation is x 1, a marginal posterior density πx 1 y) = πx 1, x 2 y)dx 2 = πx 2 )πx 1 y, x 2 )dx 2 needs to be
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate
More informationThe Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision
The Particle Filter Non-parametric implementation of Bayes filter Represents the belief (posterior) random state samples. by a set of This representation is approximate. Can represent distributions that
More informationAdaptive Monte Carlo methods
Adaptive Monte Carlo methods Jean-Michel Marin Projet Select, INRIA Futurs, Université Paris-Sud joint with Randal Douc (École Polytechnique), Arnaud Guillin (Université de Marseille) and Christian Robert
More informationMarkov chain Monte Carlo methods in atmospheric remote sensing
1 / 45 Markov chain Monte Carlo methods in atmospheric remote sensing Johanna Tamminen johanna.tamminen@fmi.fi ESA Summer School on Earth System Monitoring and Modeling July 3 Aug 11, 212, Frascati July,
More informationStat 516, Homework 1
Stat 516, Homework 1 Due date: October 7 1. Consider an urn with n distinct balls numbered 1,..., n. We sample balls from the urn with replacement. Let N be the number of draws until we encounter a ball
More informationStrong Lens Modeling (II): Statistical Methods
Strong Lens Modeling (II): Statistical Methods Chuck Keeton Rutgers, the State University of New Jersey Probability theory multiple random variables, a and b joint distribution p(a, b) conditional distribution
More informationEfficient Variational Inference in Large-Scale Bayesian Compressed Sensing
Efficient Variational Inference in Large-Scale Bayesian Compressed Sensing George Papandreou and Alan Yuille Department of Statistics University of California, Los Angeles ICCV Workshop on Information
More informationStatistical and Computational Inverse Problems with Applications Part 2: Introduction to inverse problems and example applications
Statistical and Computational Inverse Problems with Applications Part 2: Introduction to inverse problems and example applications Aku Seppänen Inverse Problems Group Department of Applied Physics University
More informationBayesian Estimation of Input Output Tables for Russia
Bayesian Estimation of Input Output Tables for Russia Oleg Lugovoy (EDF, RANE) Andrey Polbin (RANE) Vladimir Potashnikov (RANE) WIOD Conference April 24, 2012 Groningen Outline Motivation Objectives Bayesian
More informationwissen leben WWU Münster
MÜNSTER Sparsity Constraints in Bayesian Inversion Inverse Days conference in Jyväskylä, Finland. Felix Lucka 19.12.2012 MÜNSTER 2 /41 Sparsity Constraints in Inverse Problems Current trend in high dimensional
More informationCSC 2541: Bayesian Methods for Machine Learning
CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 3 More Markov Chain Monte Carlo Methods The Metropolis algorithm isn t the only way to do MCMC. We ll
More informationA quick introduction to Markov chains and Markov chain Monte Carlo (revised version)
A quick introduction to Markov chains and Markov chain Monte Carlo (revised version) Rasmus Waagepetersen Institute of Mathematical Sciences Aalborg University 1 Introduction These notes are intended to
More informationInverse Problems in the Bayesian Framework
Inverse Problems in the Bayesian Framework Daniela Calvetti Case Western Reserve University Cleveland, Ohio Raleigh, NC, July 2016 Bayes Formula Stochastic model: Two random variables X R n, B R m, where
More information18 : Advanced topics in MCMC. 1 Gibbs Sampling (Continued from the last lecture)
10-708: Probabilistic Graphical Models 10-708, Spring 2014 18 : Advanced topics in MCMC Lecturer: Eric P. Xing Scribes: Jessica Chemali, Seungwhan Moon 1 Gibbs Sampling (Continued from the last lecture)
More informationSTAT 518 Intro Student Presentation
STAT 518 Intro Student Presentation Wen Wei Loh April 11, 2013 Title of paper Radford M. Neal [1999] Bayesian Statistics, 6: 475-501, 1999 What the paper is about Regression and Classification Flexible
More informationBayesian Methods for Machine Learning
Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),
More information16 : Approximate Inference: Markov Chain Monte Carlo
10-708: Probabilistic Graphical Models 10-708, Spring 2017 16 : Approximate Inference: Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Yuan Yang, Chao-Ming Yen 1 Introduction As the target distribution
More informationPrinciples of Bayesian Inference
Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters
More informationBayesian Inference and MCMC
Bayesian Inference and MCMC Aryan Arbabi Partly based on MCMC slides from CSC412 Fall 2018 1 / 18 Bayesian Inference - Motivation Consider we have a data set D = {x 1,..., x n }. E.g each x i can be the
More informationSpatial Statistics with Image Analysis. Outline. A Statistical Approach. Johan Lindström 1. Lund October 6, 2016
Spatial Statistics Spatial Examples More Spatial Statistics with Image Analysis Johan Lindström 1 1 Mathematical Statistics Centre for Mathematical Sciences Lund University Lund October 6, 2016 Johan Lindström
More informationMonte Carlo Methods. Leon Gu CSD, CMU
Monte Carlo Methods Leon Gu CSD, CMU Approximate Inference EM: y-observed variables; x-hidden variables; θ-parameters; E-step: q(x) = p(x y, θ t 1 ) M-step: θ t = arg max E q(x) [log p(y, x θ)] θ Monte
More informationAdvances and Applications in Perfect Sampling
and Applications in Perfect Sampling Ph.D. Dissertation Defense Ulrike Schneider advisor: Jem Corcoran May 8, 2003 Department of Applied Mathematics University of Colorado Outline Introduction (1) MCMC
More informationComputer Vision Group Prof. Daniel Cremers. 14. Sampling Methods
Prof. Daniel Cremers 14. Sampling Methods Sampling Methods Sampling Methods are widely used in Computer Science as an approximation of a deterministic algorithm to represent uncertainty without a parametric
More informationMarkov Chain Monte Carlo
Markov Chain Monte Carlo Recall: To compute the expectation E ( h(y ) ) we use the approximation E(h(Y )) 1 n n h(y ) t=1 with Y (1),..., Y (n) h(y). Thus our aim is to sample Y (1),..., Y (n) from f(y).
More informationSampling Algorithms for Probabilistic Graphical models
Sampling Algorithms for Probabilistic Graphical models Vibhav Gogate University of Washington References: Chapter 12 of Probabilistic Graphical models: Principles and Techniques by Daphne Koller and Nir
More informationSAMPLING ALGORITHMS. In general. Inference in Bayesian models
SAMPLING ALGORITHMS SAMPLING ALGORITHMS In general A sampling algorithm is an algorithm that outputs samples x 1, x 2,... from a given distribution P or density p. Sampling algorithms can for example be
More informationDensity Estimation. Seungjin Choi
Density Estimation Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/
More informationMH I. Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution
MH I Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution a lot of Bayesian mehods rely on the use of MH algorithm and it s famous
More informationVariational Methods in Bayesian Deconvolution
PHYSTAT, SLAC, Stanford, California, September 8-, Variational Methods in Bayesian Deconvolution K. Zarb Adami Cavendish Laboratory, University of Cambridge, UK This paper gives an introduction to the
More informationStochastic Spectral Approaches to Bayesian Inference
Stochastic Spectral Approaches to Bayesian Inference Prof. Nathan L. Gibson Department of Mathematics Applied Mathematics and Computation Seminar March 4, 2011 Prof. Gibson (OSU) Spectral Approaches to
More informationIntroduction to Machine Learning
Introduction to Machine Learning Brown University CSCI 1950-F, Spring 2012 Prof. Erik Sudderth Lecture 25: Markov Chain Monte Carlo (MCMC) Course Review and Advanced Topics Many figures courtesy Kevin
More informationOverlapping block proposals for latent Gaussian Markov random fields
NORGES TEKNISK-NATURVITENSKAPELIGE UNIVERSITET Overlapping block proposals for latent Gaussian Markov random fields by Ingelin Steinsland and Håvard Rue PREPRINT STATISTICS NO. 8/3 NORWEGIAN UNIVERSITY
More informationMarkov Chain Monte Carlo methods
Markov Chain Monte Carlo methods By Oleg Makhnin 1 Introduction a b c M = d e f g h i 0 f(x)dx 1.1 Motivation 1.1.1 Just here Supresses numbering 1.1.2 After this 1.2 Literature 2 Method 2.1 New math As
More informationA Review of Pseudo-Marginal Markov Chain Monte Carlo
A Review of Pseudo-Marginal Markov Chain Monte Carlo Discussed by: Yizhe Zhang October 21, 2016 Outline 1 Overview 2 Paper review 3 experiment 4 conclusion Motivation & overview Notation: θ denotes the
More informationLearning the hyper-parameters. Luca Martino
Learning the hyper-parameters Luca Martino 2017 2017 1 / 28 Parameters and hyper-parameters 1. All the described methods depend on some choice of hyper-parameters... 2. For instance, do you recall λ (bandwidth
More informationPoint spread function reconstruction from the image of a sharp edge
DOE/NV/5946--49 Point spread function reconstruction from the image of a sharp edge John Bardsley, Kevin Joyce, Aaron Luttman The University of Montana National Security Technologies LLC Montana Uncertainty
More informationLecture 8: Bayesian Estimation of Parameters in State Space Models
in State Space Models March 30, 2016 Contents 1 Bayesian estimation of parameters in state space models 2 Computational methods for parameter estimation 3 Practical parameter estimation in state space
More informationMarkov chain Monte Carlo
Markov chain Monte Carlo Peter Beerli October 10, 2005 [this chapter is highly influenced by chapter 1 in Markov chain Monte Carlo in Practice, eds Gilks W. R. et al. Chapman and Hall/CRC, 1996] 1 Short
More informationECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering
ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering Lecturer: Nikolay Atanasov: natanasov@ucsd.edu Teaching Assistants: Siwei Guo: s9guo@eng.ucsd.edu Anwesan Pal:
More informationLECTURE 15 Markov chain Monte Carlo
LECTURE 15 Markov chain Monte Carlo There are many settings when posterior computation is a challenge in that one does not have a closed form expression for the posterior distribution. Markov chain Monte
More informationGraphical Models and Kernel Methods
Graphical Models and Kernel Methods Jerry Zhu Department of Computer Sciences University of Wisconsin Madison, USA MLSS June 17, 2014 1 / 123 Outline Graphical Models Probabilistic Inference Directed vs.
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is
More informationComputer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo
Group Prof. Daniel Cremers 10a. Markov Chain Monte Carlo Markov Chain Monte Carlo In high-dimensional spaces, rejection sampling and importance sampling are very inefficient An alternative is Markov Chain
More informationLecture 8: The Metropolis-Hastings Algorithm
30.10.2008 What we have seen last time: Gibbs sampler Key idea: Generate a Markov chain by updating the component of (X 1,..., X p ) in turn by drawing from the full conditionals: X (t) j Two drawbacks:
More information27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling
10-708: Probabilistic Graphical Models 10-708, Spring 2014 27 : Distributed Monte Carlo Markov Chain Lecturer: Eric P. Xing Scribes: Pengtao Xie, Khoa Luu In this scribe, we are going to review the Parallel
More informationHastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model
UNIVERSITY OF TEXAS AT SAN ANTONIO Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model Liang Jing April 2010 1 1 ABSTRACT In this paper, common MCMC algorithms are introduced
More informationAdaptive Posterior Approximation within MCMC
Adaptive Posterior Approximation within MCMC Tiangang Cui (MIT) Colin Fox (University of Otago) Mike O Sullivan (University of Auckland) Youssef Marzouk (MIT) Karen Willcox (MIT) 06/January/2012 C, F,
More informationStat 451 Lecture Notes Markov Chain Monte Carlo. Ryan Martin UIC
Stat 451 Lecture Notes 07 12 Markov Chain Monte Carlo Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapters 8 9 in Givens & Hoeting, Chapters 25 27 in Lange 2 Updated: April 4, 2016 1 / 42 Outline
More informationBayesian Inference in Astronomy & Astrophysics A Short Course
Bayesian Inference in Astronomy & Astrophysics A Short Course Tom Loredo Dept. of Astronomy, Cornell University p.1/37 Five Lectures Overview of Bayesian Inference From Gaussians to Periodograms Learning
More informationResults: MCMC Dancers, q=10, n=500
Motivation Sampling Methods for Bayesian Inference How to track many INTERACTING targets? A Tutorial Frank Dellaert Results: MCMC Dancers, q=10, n=500 1 Probabilistic Topological Maps Results Real-Time
More informationComputer Vision Group Prof. Daniel Cremers. 11. Sampling Methods: Markov Chain Monte Carlo
Group Prof. Daniel Cremers 11. Sampling Methods: Markov Chain Monte Carlo Markov Chain Monte Carlo In high-dimensional spaces, rejection sampling and importance sampling are very inefficient An alternative
More information17 : Markov Chain Monte Carlo
10-708: Probabilistic Graphical Models, Spring 2015 17 : Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Heran Lin, Bin Deng, Yun Huang 1 Review of Monte Carlo Methods 1.1 Overview Monte Carlo
More informationBayesian Image Segmentation Using MRF s Combined with Hierarchical Prior Models
Bayesian Image Segmentation Using MRF s Combined with Hierarchical Prior Models Kohta Aoki 1 and Hiroshi Nagahashi 2 1 Interdisciplinary Graduate School of Science and Engineering, Tokyo Institute of Technology
More informationBayesian Inference for DSGE Models. Lawrence J. Christiano
Bayesian Inference for DSGE Models Lawrence J. Christiano Outline State space-observer form. convenient for model estimation and many other things. Bayesian inference Bayes rule. Monte Carlo integation.
More informationMARKOV CHAIN MONTE CARLO
MARKOV CHAIN MONTE CARLO RYAN WANG Abstract. This paper gives a brief introduction to Markov Chain Monte Carlo methods, which offer a general framework for calculating difficult integrals. We start with
More informationBayesian rules of probability as principles of logic [Cox] Notation: pr(x I) is the probability (or pdf) of x being true given information I
Bayesian rules of probability as principles of logic [Cox] Notation: pr(x I) is the probability (or pdf) of x being true given information I 1 Sum rule: If set {x i } is exhaustive and exclusive, pr(x
More informationComputer Intensive Methods in Mathematical Statistics
Computer Intensive Methods in Mathematical Statistics Department of mathematics johawes@kth.se Lecture 16 Advanced topics in computational statistics 18 May 2017 Computer Intensive Methods (1) Plan of
More informationST 740: Markov Chain Monte Carlo
ST 740: Markov Chain Monte Carlo Alyson Wilson Department of Statistics North Carolina State University October 14, 2012 A. Wilson (NCSU Stsatistics) MCMC October 14, 2012 1 / 20 Convergence Diagnostics:
More informationMCMC algorithms for fitting Bayesian models
MCMC algorithms for fitting Bayesian models p. 1/1 MCMC algorithms for fitting Bayesian models Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota MCMC algorithms for fitting Bayesian models
More informationF denotes cumulative density. denotes probability density function; (.)
BAYESIAN ANALYSIS: FOREWORDS Notation. System means the real thing and a model is an assumed mathematical form for the system.. he probability model class M contains the set of the all admissible models
More informationMonte Carlo simulation inspired by computational optimization. Colin Fox Al Parker, John Bardsley MCQMC Feb 2012, Sydney
Monte Carlo simulation inspired by computational optimization Colin Fox fox@physics.otago.ac.nz Al Parker, John Bardsley MCQMC Feb 2012, Sydney Sampling from π(x) Consider : x is high-dimensional (10 4
More informationRiemann Manifold Methods in Bayesian Statistics
Ricardo Ehlers ehlers@icmc.usp.br Applied Maths and Stats University of São Paulo, Brazil Working Group in Statistical Learning University College Dublin September 2015 Bayesian inference is based on Bayes
More informationPart 8: GLMs and Hierarchical LMs and GLMs
Part 8: GLMs and Hierarchical LMs and GLMs 1 Example: Song sparrow reproductive success Arcese et al., (1992) provide data on a sample from a population of 52 female song sparrows studied over the course
More informationDiscretization-invariant Bayesian inversion and Besov space priors. Samuli Siltanen RICAM Tampere University of Technology
Discretization-invariant Bayesian inversion and Besov space priors Samuli Siltanen RICAM 28.10.2008 Tampere University of Technology http://math.tkk.fi/inverse-coe/ This is a joint work with Matti Lassas
More informationLecture 4: Probabilistic Learning
DD2431 Autumn, 2015 1 Maximum Likelihood Methods Maximum A Posteriori Methods Bayesian methods 2 Classification vs Clustering Heuristic Example: K-means Expectation Maximization 3 Maximum Likelihood Methods
More informationGentle Introduction to Infinite Gaussian Mixture Modeling
Gentle Introduction to Infinite Gaussian Mixture Modeling with an application in neuroscience By Frank Wood Rasmussen, NIPS 1999 Neuroscience Application: Spike Sorting Important in neuroscience and for
More informationEmpirical Bayes Unfolding of Elementary Particle Spectra at the Large Hadron Collider
Empirical Bayes Unfolding of Elementary Particle Spectra at the Large Hadron Collider Mikael Kuusela Institute of Mathematics, EPFL Statistics Seminar, University of Bristol June 13, 2014 Joint work with
More informationMetropolis Hastings. Rebecca C. Steorts Bayesian Methods and Modern Statistics: STA 360/601. Module 9
Metropolis Hastings Rebecca C. Steorts Bayesian Methods and Modern Statistics: STA 360/601 Module 9 1 The Metropolis-Hastings algorithm is a general term for a family of Markov chain simulation methods
More informationMonte Carlo methods for sampling-based Stochastic Optimization
Monte Carlo methods for sampling-based Stochastic Optimization Gersende FORT LTCI CNRS & Telecom ParisTech Paris, France Joint works with B. Jourdain, T. Lelièvre, G. Stoltz from ENPC and E. Kuhn from
More informationLecture: Gaussian Process Regression. STAT 6474 Instructor: Hongxiao Zhu
Lecture: Gaussian Process Regression STAT 6474 Instructor: Hongxiao Zhu Motivation Reference: Marc Deisenroth s tutorial on Robot Learning. 2 Fast Learning for Autonomous Robots with Gaussian Processes
More informationPattern Recognition and Machine Learning
Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability
More informationAdvanced Machine Learning
Advanced Machine Learning Nonparametric Bayesian Models --Learning/Reasoning in Open Possible Worlds Eric Xing Lecture 7, August 4, 2009 Reading: Eric Xing Eric Xing @ CMU, 2006-2009 Clustering Eric Xing
More informationMarkov Chain Monte Carlo methods
Markov Chain Monte Carlo methods Tomas McKelvey and Lennart Svensson Signal Processing Group Department of Signals and Systems Chalmers University of Technology, Sweden November 26, 2012 Today s learning
More informationWinter 2019 Math 106 Topics in Applied Mathematics. Lecture 9: Markov Chain Monte Carlo
Winter 2019 Math 106 Topics in Applied Mathematics Data-driven Uncertainty Quantification Yoonsang Lee (yoonsang.lee@dartmouth.edu) Lecture 9: Markov Chain Monte Carlo 9.1 Markov Chain A Markov Chain Monte
More informationIntroduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016
Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016 EPSY 905: Intro to Bayesian and MCMC Today s Class An
More informationComputer Vision Group Prof. Daniel Cremers. 11. Sampling Methods
Prof. Daniel Cremers 11. Sampling Methods Sampling Methods Sampling Methods are widely used in Computer Science as an approximation of a deterministic algorithm to represent uncertainty without a parametric
More information(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis
Summarizing a posterior Given the data and prior the posterior is determined Summarizing the posterior gives parameter estimates, intervals, and hypothesis tests Most of these computations are integrals
More informationMarkov Random Fields
Markov Random Fields Umamahesh Srinivas ipal Group Meeting February 25, 2011 Outline 1 Basic graph-theoretic concepts 2 Markov chain 3 Markov random field (MRF) 4 Gauss-Markov random field (GMRF), and
More informationMarkov Chain Monte Carlo Using the Ratio-of-Uniforms Transformation. Luke Tierney Department of Statistics & Actuarial Science University of Iowa
Markov Chain Monte Carlo Using the Ratio-of-Uniforms Transformation Luke Tierney Department of Statistics & Actuarial Science University of Iowa Basic Ratio of Uniforms Method Introduced by Kinderman and
More informationMarkov Chain Monte Carlo, Numerical Integration
Markov Chain Monte Carlo, Numerical Integration (See Statistics) Trevor Gallen Fall 2015 1 / 1 Agenda Numerical Integration: MCMC methods Estimating Markov Chains Estimating latent variables 2 / 1 Numerical
More informationGeneral Construction of Irreversible Kernel in Markov Chain Monte Carlo
General Construction of Irreversible Kernel in Markov Chain Monte Carlo Metropolis heat bath Suwa Todo Department of Applied Physics, The University of Tokyo Department of Physics, Boston University (from
More informationSession 3A: Markov chain Monte Carlo (MCMC)
Session 3A: Markov chain Monte Carlo (MCMC) John Geweke Bayesian Econometrics and its Applications August 15, 2012 ohn Geweke Bayesian Econometrics and its Session Applications 3A: Markov () chain Monte
More informationLinear Dynamical Systems
Linear Dynamical Systems Sargur N. srihari@cedar.buffalo.edu Machine Learning Course: http://www.cedar.buffalo.edu/~srihari/cse574/index.html Two Models Described by Same Graph Latent variables Observations
More informationCosmoSIS Webinar Part 1
CosmoSIS Webinar Part 1 An introduction to cosmological parameter estimation and sampling Presenting for the CosmoSIS team: Elise Jennings (Fermilab, KICP) Joe Zuntz (University of Manchester) Vinicius
More informationDynamic System Identification using HDMR-Bayesian Technique
Dynamic System Identification using HDMR-Bayesian Technique *Shereena O A 1) and Dr. B N Rao 2) 1), 2) Department of Civil Engineering, IIT Madras, Chennai 600036, Tamil Nadu, India 1) ce14d020@smail.iitm.ac.in
More informationBayesian model selection in graphs by using BDgraph package
Bayesian model selection in graphs by using BDgraph package A. Mohammadi and E. Wit March 26, 2013 MOTIVATION Flow cytometry data with 11 proteins from Sachs et al. (2005) RESULT FOR CELL SIGNALING DATA
More informationComputer Practical: Metropolis-Hastings-based MCMC
Computer Practical: Metropolis-Hastings-based MCMC Andrea Arnold and Franz Hamilton North Carolina State University July 30, 2016 A. Arnold / F. Hamilton (NCSU) MH-based MCMC July 30, 2016 1 / 19 Markov
More informationeqr094: Hierarchical MCMC for Bayesian System Reliability
eqr094: Hierarchical MCMC for Bayesian System Reliability Alyson G. Wilson Statistical Sciences Group, Los Alamos National Laboratory P.O. Box 1663, MS F600 Los Alamos, NM 87545 USA Phone: 505-667-9167
More information