Monte Carlo importance sampling and Markov chain

Similar documents
Monte Carlo, importance sampling through Markov chain and simulated annealing

Monte Carlo and cold gases. Lode Pollet.

Introduction to Machine Learning CMU-10701

Markov Chains and MCMC

The Monte Carlo Method

Simulation - Lectures - Part III Markov chain Monte Carlo

Math 456: Mathematical Modeling. Tuesday, April 9th, 2018

(Classical) Information Theory II: Source coding

Convergence Rate of Markov Chains

Reminder of some Markov Chain properties:

Advanced Sampling Algorithms

Markov Chain Monte Carlo Inference. Siamak Ravanbakhsh Winter 2018

Random Walks A&T and F&S 3.1.2

Math 456: Mathematical Modeling. Tuesday, March 6th, 2018

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods: Markov Chain Monte Carlo

Markov Chains. X(t) is a Markov Process if, for arbitrary times t 1 < t 2 <... < t k < t k+1. If X(t) is discrete-valued. If X(t) is continuous-valued

Introduction to MCMC. DB Breakfast 09/30/2011 Guozhang Wang

Convex Optimization CMU-10725

Quantifying Uncertainty

Lecture 7. We can regard (p(i, j)) as defining a (maybe infinite) matrix P. Then a basic fact is

6 Markov Chain Monte Carlo (MCMC)

Lecture 6: Markov Chain Monte Carlo

Monte Carlo Methods. Leon Gu CSD, CMU

Machine Learning for Data Science (CS4786) Lecture 24

Metropolis/Variational Monte Carlo. Microscopic Theories of Nuclear Structure, Dynamics and Electroweak Currents June 12-30, 2017, ECT*, Trento, Italy

Stat 535 C - Statistical Computing & Monte Carlo Methods. Lecture February Arnaud Doucet

Law of large numbers for Markov chains

Numerical methods for lattice field theory

= P{X 0. = i} (1) If the MC has stationary transition probabilities then, = i} = P{X n+1

Robert Collins CSE586, PSU. Markov-Chain Monte Carlo

Introduction to Machine Learning CMU-10701

MCMC: Markov Chain Monte Carlo

CSC 446 Notes: Lecture 13

Homework set 2 - Solutions

Monte Carlo (MC) Simulation Methods. Elisa Fadda

7.1 Coupling from the Past

Monte Carlo Methods in Statistical Mechanics

Stat 516, Homework 1

Markov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains

1 Probabilities. 1.1 Basics 1 PROBABILITIES

Parallel Tempering Algorithm in Monte Carlo Simulation

Positive and null recurrent-branching Process

A quick introduction to Markov chains and Markov chain Monte Carlo (revised version)

Markov chain Monte Carlo Lecture 9

On Markov Chain Monte Carlo

Markov Chain Monte Carlo Methods

Monte Caro simulations

Minicourse on: Markov Chain Monte Carlo: Simulation Techniques in Statistics

MATH 56A: STOCHASTIC PROCESSES CHAPTER 7

Bayesian Inference and MCMC

17 : Markov Chain Monte Carlo

1 Probabilities. 1.1 Basics 1 PROBABILITIES

Markov Chain Monte Carlo The Metropolis-Hastings Algorithm

Winter 2019 Math 106 Topics in Applied Mathematics. Lecture 9: Markov Chain Monte Carlo

Markov Processes. Stochastic process. Markov process

16 : Approximate Inference: Markov Chain Monte Carlo

Advanced Monte Carlo Methods Problems

Monte Carlo. Lecture 15 4/9/18. Harvard SEAS AP 275 Atomistic Modeling of Materials Boris Kozinsky

Markov chain Monte Carlo

CSC 2541: Bayesian Methods for Machine Learning

Introduc)on to Bayesian Methods

Adaptive Monte Carlo methods

Computer Vision Group Prof. Daniel Cremers. 14. Sampling Methods

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods

Pattern Recognition and Machine Learning. Bishop Chapter 11: Sampling Methods

Interlude: Practice Final

16 : Markov Chain Monte Carlo (MCMC)

Bayesian Methods with Monte Carlo Markov Chains II

Random Target Lemma. Sucharita Jayanti. February 21, 2013

Markov Chain Monte Carlo. Simulated Annealing.

A Monte Carlo Implementation of the Ising Model in Python

MCMC and Gibbs Sampling. Sargur Srihari

Lecture 2 : CS6205 Advanced Modeling and Simulation

Data Analysis I. Dr Martin Hendry, Dept of Physics and Astronomy University of Glasgow, UK. 10 lectures, beginning October 2006

Quantum Monte Carlo. Matthias Troyer, ETH Zürich

References. Markov-Chain Monte Carlo. Recall: Sampling Motivation. Problem. Recall: Sampling Methods. CSE586 Computer Vision II

Stochastic Simulation

Probability & Computing

Figure 10.1: Recording when the event E occurs

Artificial Intelligence

Simulation of the two-dimensional square-lattice Lenz-Ising model in Python

Chapter 10. Optimization Simulated annealing

Markov Chain Monte Carlo

Markov and Gibbs Random Fields

Minimal basis for connected Markov chain over 3 3 K contingency tables with fixed two-dimensional marginals. Satoshi AOKI and Akimichi TAKEMURA

The Metropolis-Hastings Algorithm. June 8, 2012

Markov Chain Monte Carlo

Likelihood Inference for Lattice Spatial Processes

MCMC notes by Mark Holder

Markov Chain Monte Carlo Method

Markov-Chain Monte Carlo

Physics 115/242 Monte Carlo simulations in Statistical Physics

Lectures on Stochastic Stability. Sergey FOSS. Heriot-Watt University. Lecture 4. Coupling and Harris Processes

The Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision

Markov Chain Monte Carlo

Example: physical systems. If the state space. Example: speech recognition. Context can be. Example: epidemics. Suppose each infected

CS 361: Probability & Statistics

LECTURE 3. Last time:

Markov Chain Monte Carlo, Numerical Integration

Markov chain Monte Carlo

Transcription:

Monte Carlo importance sampling and Markov chain If a configuration in phase space is denoted by X, the probability for configuration according to Boltzman is ρ(x) e βe(x) β = 1 T (1) How to sample over the whole phase space for a general problem? How to generate configurations? Brute force: generate a truly random configuration X and accept it with probability e βe(x) where all E > 0. Successive X are statistically independent. VERY INEFFICIENT Markov chain: Successive configurations X i, X i+1 are NOT statistically independent but are distributed according to Boltzman distribution. What is the difference between Markov chain and uncorrelated sequence? Truly random or uncorrelated sequence of configurations satisfies the identity P(X 1, X 2,,P XN ) = P 1 (X 1 )P 1 (X 2 ) P 1 (X N )

Markov chain satisfies the equation P(X 1, X 2,,P XN ) = P 1 (X 1 )T(X 1 X 2 )T(X 2 X 3 ) T(X N 1 X N ) where the transition probabilities T(X X ) are normalized X T(X X ) = 1 We want to generate Markov chain where distribution of states is proportional to e βe(x) and this distribution should be independent of the position within the chain and independent of the initial configuration. The necessary conditions for generating such Markov chain is that every configuration in phase space should be accesible from any other configuration within finite number of steps (connectedness or irreducibility) - (Be careful to check this condition when choosing Monte Carlo step!) We need to find transition probability T(X X ) which leads to a given stationary distribution ρ(x) (in this case ρ(x) e βe(x) ).

The probability for X decreases, if system goes from X to any other X : X ρ(x)t(x X ) and increases if X configuration is visited from any other state X : X ρ(x )T(X X). The change of probability for X is therefore ρ(x, t + 1) ρ(x, t) = ρ(x)t(x X ) + ρ(x )T(X X) (2) X X We look for stationary solution, i.e., ρ(x, t + 1) ρ(x, t) = 0 and therefore ρ(x)t(x X ) = ρ(x )T(X X) (3) X X General solution of this equation is not accesible, but a particular solution is obvious ρ(x)t(x X ) = ρ(x )T(X X) (4) This solution is called DETAIL BALANCE solution.

To construct algorithm, we devide transition prob. T(X X ) = ω XX A XX : trial step probability ω XX which is symmetric, i.e., ω XX = ω X X (for example spin flip in ising: ω XX is 1/L 2 if X and X differ for a single spin flip and zero otherwise ) and acceptance probability A XX (for example accepting of rejecting new configuration with probability proportional to min(1, exp( β(e(x ) E(X))))). Detail balance condition becomes ρ(x ) ρ(x) = A XX A X X Metropolis chooses A XX = 1 A XX = ρ(x ) ρ(x) if ρ(x ) > ρ(x) if ρ(x ) < ρ(x). Obviously, this acceptance probability satisfies detail balance condition and therefore leads to desired Markov chain with stationary probability for any configuration X ρ(x) for long times. (5)

To summarize Metropolis algorithm T(X X ) = ω XX A XX X ω XX = 1; ω XX = ω X X ω XX > 0 for all X, X after finite number of steps A XX = min(1, ρ(x ) ρ(x) ) How to accept a step with probability A XX < 1? One can generate a random number r [0, 1] and accept the step if r < A XX. Keep in mind: Configurations that are generated by Markov chain are correlated. The theory guarantees that we arrive at invariant distribution ρ for long times. Two configurations are statistically independent only if they are far apart in the Markov chain. This distance is called correlation time (Be careful: To meassure distance in Markov chain, every step counts, not only successful.)

The average of any quantity can be calculated as usual A = where n 0 steps are used to warm-up. 1 n n 0 n i>n 0 A i The error of the quantity, however, is much bigger than the following quantity 1 n n 0 n i>n 0 (A i A) 2 Imagine the extreme limit of correlations when all values A i are the same. We would estimate that standard deviation is zero regardless of the actual error! To compute standard deviation, we need to group meassurements within the correlation time into bins and than estimate the standard deviation of the bins: B l = 1 i<n l +N 0 N 0 i=n l A i (6)

σ 2 = 1 M (B j A) 2 (7) M 1 j=0 where we took into account that A = B. The correlation time (here denoted by N 0 ) is not very easy to estimate. Maybe the best algorithm is to compute σ 2 for few different N 0 and as long as σ 2 is increasing with N 0, the correlation time is still larger than N 0. When σ 2 stops changing with increasing N 0, we reached correlation time and σ 2 is a good estimation of standard deviation.