Introduction to MCMC. DB Breakfast 09/30/2011 Guozhang Wang

Similar documents
16 : Markov Chain Monte Carlo (MCMC)

Bayesian Inference and MCMC

Markov Chains and MCMC

Introduction to Machine Learning CMU-10701

STA 4273H: Statistical Machine Learning

17 : Markov Chain Monte Carlo

Lecture 7 and 8: Markov Chain Monte Carlo

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods

CSC 446 Notes: Lecture 13

Computer Vision Group Prof. Daniel Cremers. 14. Sampling Methods

Answers and expectations

Quantifying Uncertainty

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning

The Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision

Monte Carlo Methods. Leon Gu CSD, CMU

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods: Markov Chain Monte Carlo

CPSC 540: Machine Learning


Markov chain Monte Carlo

16 : Approximate Inference: Markov Chain Monte Carlo

MCMC: Markov Chain Monte Carlo

CS242: Probabilistic Graphical Models Lecture 7B: Markov Chain Monte Carlo & Gibbs Sampling

StarAI Full, 6+1 pages Short, 2 page position paper or abstract

Markov Chain Monte Carlo Inference. Siamak Ravanbakhsh Winter 2018

Markov chain Monte Carlo

MCMC and Gibbs Sampling. Kayhan Batmanghelich

Lecture 6: Markov Chain Monte Carlo

18 : Advanced topics in MCMC. 1 Gibbs Sampling (Continued from the last lecture)

MCMC and Gibbs Sampling. Sargur Srihari

SAMPLING ALGORITHMS. In general. Inference in Bayesian models

Markov chain Monte Carlo

19 : Slice Sampling and HMC

Machine Learning for Data Science (CS4786) Lecture 24

Reminder of some Markov Chain properties:

References. Markov-Chain Monte Carlo. Recall: Sampling Motivation. Problem. Recall: Sampling Methods. CSE586 Computer Vision II

Stochastic optimization Markov Chain Monte Carlo

Introduction to Computational Biology Lecture # 14: MCMC - Markov Chain Monte Carlo

Generating the Sample

Markov Chain Monte Carlo (MCMC)

Sampling Methods (11/30/04)

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo

Lecture 8: The Metropolis-Hastings Algorithm

Computational statistics

Markov-Chain Monte Carlo

Statistical Machine Learning Lecture 8: Markov Chain Monte Carlo Sampling

14 : Approximate Inference Monte Carlo Methods

Markov Chain Monte Carlo (MCMC)

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Markov Chain Monte Carlo methods

MCMC Methods: Gibbs and Metropolis

Convex Optimization CMU-10725

Sampling Algorithms for Probabilistic Graphical models

Markov Chains and MCMC

Markov Chain Monte Carlo

CS281A/Stat241A Lecture 22

Bayesian Methods and Uncertainty Quantification for Nonlinear Inverse Problems

Probabilistic Graphical Models Lecture 17: Markov chain Monte Carlo

Markov chain Monte Carlo methods in atmospheric remote sensing

MARKOV CHAIN MONTE CARLO

Markov chain Monte Carlo

STA 294: Stochastic Processes & Bayesian Nonparametrics

Graphical Models and Kernel Methods

A quick introduction to Markov chains and Markov chain Monte Carlo (revised version)

ELEC633: Graphical Models

The Ising model and Markov chain Monte Carlo

Bayesian Methods for Machine Learning

6 Markov Chain Monte Carlo (MCMC)

Monte Carlo importance sampling and Markov chain

Probabilistic Machine Learning

Robert Collins CSE586, PSU Intro to Sampling Methods

Markov Chain Monte Carlo The Metropolis-Hastings Algorithm

Markov chain Monte Carlo Lecture 9

Bayesian Phylogenetics:

Introduction to Machine Learning CMU-10701

MSc MT15. Further Statistical Methods: MCMC. Lecture 5-6: Markov chains; Metropolis Hastings MCMC. Notes and Practicals available at

Markov Chain Monte Carlo Methods for Stochastic Optimization

Bayesian Methods with Monte Carlo Markov Chains II

Particle Filtering Approaches for Dynamic Stochastic Optimization

Brief introduction to Markov Chain Monte Carlo

Stat 451 Lecture Notes Markov Chain Monte Carlo. Ryan Martin UIC

Monte Carlo Inference Methods

Pattern Recognition and Machine Learning. Bishop Chapter 11: Sampling Methods

LECTURE 15 Markov chain Monte Carlo

Who was Bayes? Bayesian Phylogenetics. What is Bayes Theorem?

Adaptive Monte Carlo methods

STAT 425: Introduction to Bayesian Analysis

Bayesian Phylogenetics

Stat 535 C - Statistical Computing & Monte Carlo Methods. Lecture February Arnaud Doucet

Markov Chain Monte Carlo methods

Recap. Probability, stochastic processes, Markov chains. ELEC-C7210 Modeling and analysis of communication networks

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet.

Stochastic Simulation

Minicourse on: Markov Chain Monte Carlo: Simulation Techniques in Statistics

Markov Chain Monte Carlo

Need for Sampling in Machine Learning. Sargur Srihari

Algorithms for Reasoning with Probabilistic Graphical Models. Class 3: Approximate Inference. International Summer School on Deep Learning July 2017

an introduction to bayesian inference

MCMC algorithms for fitting Bayesian models

Computer intensive statistical methods

Monte Carlo Methods. Geoff Gordon February 9, 2006

Transcription:

Introduction to MCMC DB Breakfast 09/30/2011 Guozhang Wang

Motivation: Statistical Inference Joint Distribution Sleeps Well Playground Sunny Bike Ride Pleasant dinner Productive day Posterior Estimation Graphical Models

Motivation: Statistical Physics Energy Model Ising Model Thermal Eqm. Estimation

Problem I: Integral Computation Posterior Estimation: Thermal Eqm. Estimation:

Problem I Rewrite: Sampling Generate samples {x (r) } R from the probability distribution p(x). If we can solve this problem, we can solve the R ( r ) ( r ) integral computation by: f ( x ) p ( x ) We will show later this estimator is unbiased with very nice variance bound i

Deterministic Methods Numerical Integration Choose fixed points in the distribution Use their probability values Unbiased, but the variance is exponential to dimension

Random Methods: Monte Carlo Generate samples i.i.d Compute samples probability Approximate integral by samples integration

Merits of Monte Carlo Law of Large Numbers Function f(x) over random variable x I.i.d random samples drawn from p(x) 1 n f ( X ) f ( x) p ( x dx as n 1 n i i ) Central Limit Theorem I.i.d samples with expectation μ and variance σ 2 Sample distribution normal(μ, σ 2 /n) Variance Not Depend on Dimension!

Simple Sampling Complex distributions Known CDF: inversion methods Simpler q(x) : Rejection sampling Can compute density: importance sampling

Come Back to Statistical Inference Forward Sampling Repeated sample x F (i), x R (i), x E (i) based on prior and conditionals Discard x (i) when x E (i) is not observed x E When N samples retained, estimate p(x F x E ) as Problem: low acceptance rate

Problem II: Curse of Dimensionality The prob. dense area shrinks as dimension d arises Harder to sample in this area to get enough information of the distribution Acceptance rate decreases exponentially with d

Solution: Sampling with Guide Avoid random-walk, but sample variables conditional on previous samples Note: violate the i.i.d condition of LLN and CLT

Markov Chain Memoryless Random Process Transition probability A: p(x t+1 ) = A*p(x t ) Non-independent Samples, thus no guarantee of convergence

Mission Impossible? How can we set the transition probabilities such that the 1) there is a equilibrium, and 2) equilibrium distribution is the target distribution, without knowing what the target is?

Markov Chain Properties A Markov chain is called: Stationary, if there exists P such that P = A*P; note that multiple stationary distribution can exist. Aperiodic, if there is no cycles with transition probability 1. Irreducible, if has positive probability of reaching any state from any other Non-transient, if it can always return to a state after visiting it Reversible w.r.t P, if P(x=i) A[ij] = P(x=j) A[ji]

Convergence of Markov Chain If the chain is Reversible w.r.t. P, then P is its stationary distribution. And, if the chain is Aperiodic and Irreducible, it have a single stationary distribution, which it will converge to almost surely. And, if the chain is Non-transient, it will always converge to its stationary distribution from any starting states. Goal: Design alg. to satisfy all these properties.

Metropolis-Hastings

MCDB: A Monte Carlo Approach to Managing Uncertain Data Used for probabilistic Data management, where uncertainty can be expressed via distribution function. CREATE TABLE SBP DATA(PID, GENDER, SBP) AS FOR EACH p in PATIENTS WITH SBP AS Normal ( (SELECT s.mean, s.std FROM SPB PARAM s)) SELECT p.pid, p.gender, b.value FROM SBP b

MCDB: A Monte Carlo Approach to Managing Uncertain Data Query processing Sample instances from the distribution function Execute the query on each sampled DB instance, thereby approximate the query-result distribution Use Monte Carlo properties to compute mean, variance, quantiles, etc. Some optimization Tricks Tuple bundles Split and merge

Limits MCDB: A Monte Carlo Approach to Managing Uncertain Data Risk analysis concerns with quintiles mostly Requires lots of samples to bound error Actually is the curse of dimensionality MCDB-R: Risk Analysis in the Database Monte Carlo + Markov Chain (MCMC) Use Gibbs sampling

Thanks!