A =, where d n w = dw

Similar documents
3.320 Lecture 18 (4/12/05)

Random Walks A&T and F&S 3.1.2

Path Integrals. Andreas Wipf Theoretisch-Physikalisches-Institut Friedrich-Schiller-Universität, Max Wien Platz Jena

Markov Processes. Stochastic process. Markov process

Copyright 2001 University of Cambridge. Not to be quoted or copied without permission.

Advanced Monte Carlo Methods Problems

Physics 115/242 Monte Carlo simulations in Statistical Physics

Monte Carlo. Lecture 15 4/9/18. Harvard SEAS AP 275 Atomistic Modeling of Materials Boris Kozinsky

CS 781 Lecture 9 March 10, 2011 Topics: Local Search and Optimization Metropolis Algorithm Greedy Optimization Hopfield Networks Max Cut Problem Nash

CHAPTER 4. Cluster expansions

Finite-Horizon Statistics for Markov chains

Markov Chain Monte Carlo The Metropolis-Hastings Algorithm

Monte Carlo (MC) Simulation Methods. Elisa Fadda

Markov Chain Monte Carlo Simulations and Their Statistical Analysis An Overview

Section 4.4 Reduction to Symmetric Tridiagonal Form

Project: Vibration of Diatomic Molecules

Monte Caro simulations

Time-dependent linear-response variational Monte Carlo.

NATIONAL UNIVERSITY OF SINGAPORE PC5215 NUMERICAL RECIPES WITH APPLICATIONS. (Semester I: AY ) Time Allowed: 2 Hours

lattice QCD and the hadron spectrum Jozef Dudek ODU/JLab

Monte Carlo Methods in Statistical Mechanics

5. Simulated Annealing 5.1 Basic Concepts. Fall 2010 Instructor: Dr. Masoud Yaghini

Introduction to Machine Learning CMU-10701

Monte Carlo Simulation of the Ising Model. Abstract

Methods of Fermion Monte Carlo

Lecture 10. The Dirac equation. WS2010/11: Introduction to Nuclear and Particle Physics

Understanding Molecular Simulation 2009 Monte Carlo and Molecular Dynamics in different ensembles. Srikanth Sastry

Physics 127b: Statistical Mechanics. Renormalization Group: 1d Ising Model. Perturbation expansion

Tensor network simulations of strongly correlated quantum systems

6. Auxiliary field continuous time quantum Monte Carlo

Symmetries, Groups, and Conservation Laws

I. QUANTUM MONTE CARLO METHODS: INTRODUCTION AND BASICS

Lattice spin models: Crash course

Applicability of Quasi-Monte Carlo for lattice systems

ICCP Project 2 - Advanced Monte Carlo Methods Choose one of the three options below

Matrix balancing and robust Monte Carlo algorithm for evaluating dominant eigenpair

The Phase Transition of the 2D-Ising Model

Regularization Physics 230A, Spring 2007, Hitoshi Murayama

Topological defects and its role in the phase transition of a dense defect system

Kinetic Monte Carlo (KMC)

Clusters and Percolation

lattice QCD and the hadron spectrum Jozef Dudek ODU/JLab

A Monte Carlo Implementation of the Ising Model in Python

5 Irreducible representations

8.334: Statistical Mechanics II Problem Set # 4 Due: 4/9/14 Transfer Matrices & Position space renormalization

Energy Landscapes of Deep Networks

Need for Deep Networks Perceptron. Can only model linear functions. Kernel Machines. Non-linearity provided by kernels

LS.2 Homogeneous Linear Systems with Constant Coefficients

Lecture 11: Non-Equilibrium Statistical Physics

Cluster Algorithms to Reduce Critical Slowing Down

Numerical methods for lattice field theory

Kinetic Monte Carlo (KMC) Kinetic Monte Carlo (KMC)

Simulation of the two-dimensional square-lattice Lenz-Ising model in Python

In this section again we shall assume that the matrix A is m m, real and symmetric.

Computational Physics. J. M. Thijssen

Data Mining and Analysis: Fundamental Concepts and Algorithms

3.320: Lecture 19 (4/14/05) Free Energies and physical Coarse-graining. ,T) + < σ > dµ

Lecture 12: Randomized Least-squares Approximation in Practice, Cont. 12 Randomized Least-squares Approximation in Practice, Cont.

Any live cell with less than 2 live neighbours dies. Any live cell with 2 or 3 live neighbours lives on to the next step.

The quantum state as a vector

6 Markov Chain Monte Carlo (MCMC)

+ V = 0, j = 1,..., 3N (7.1) i =1, 3 m = 1, N, X mi, V X mi. i = 0 in the equilibrium. X mi X nj

An indicator for the number of clusters using a linear map to simplex structure

QM and Angular Momentum

A simple technique to estimate partition functions and equilibrium constants from Monte Carlo simulations

An Alternative Proof of Primitivity of Indecomposable Nonnegative Matrices with a Positive Trace

Parton shower evolution with subleading color

Linearization of Differential Equation Models

Symmetries for fun and profit

Chapter 2 Approximation Methods Can be Used When Exact Solutions to the Schrödinger Equation Can Not be Found.

2 Eigenvectors and Eigenvalues in abstract spaces.

Their Statistical Analvsis. With Web-Based Fortran Code. Berg

4. Cluster update algorithms

Math 1553, Introduction to Linear Algebra

Some Notes on Linear Algebra

12. LOCAL SEARCH. gradient descent Metropolis algorithm Hopfield neural networks maximum cut Nash equilibria

Newtonian Mechanics. Chapter Classical space-time

Multiple time step Monte Carlo

EIGENVALUES AND EIGENVECTORS 3

THE DIRAC EQUATION (A REVIEW) We will try to find the relativistic wave equation for a particle.

5.3 The Power Method Approximation of the Eigenvalue of Largest Module

Vacuum Energy and Effective Potentials

σ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) =

On the exact bit error probability for Viterbi decoding of convolutional codes

1 Geometry of R Conic Sections Parametric Equations More Parametric Equations Polar Coordinates...

Numerical Studies of Adiabatic Quantum Computation applied to Optimization and Graph Isomorphism

Classical Monte Carlo Simulations

Numerical Analysis Lecture Notes

J ij S i S j B i S i (1)

22.3. Repeated Eigenvalues and Symmetric Matrices. Introduction. Prerequisites. Learning Outcomes

Quantum wires, orthogonal polynomials and Diophantine approximation

Chem 3502/4502 Physical Chemistry II (Quantum Mechanics) 3 Credits Spring Semester 2006 Christopher J. Cramer. Lecture 9, February 8, 2006

A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring

CS 246 Review of Linear Algebra 01/17/19

Jim Lambers MAT 610 Summer Session Lecture 2 Notes

Large-N Quantum Field Theories and Nonlinear Random Processes

Numerical Methods in Quantum Field Theories

Time-Independent Perturbation Theory

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods: Markov Chain Monte Carlo

The complex projective line

Transcription:

Chapter 9 Monte Carlo Simulations So far we have either dealt with exactly soluble systems like the (time-dependent) oscillator or with various expansions like the semiclassical, perturbative and high-temperature expansion. In this section I introduce a numerical method which plays an important role in modern developments in (gauge-) field theories - the Monte-Carlo simulations. One conveniently starts from the lattice approximation to Z(β,q) = K( hβ,q,q) in (6.2) for the density of the partition function. In the finite dimensional integral over the points w,...,w n defining the broken line path one sets w 0 = w n = q and integrates over q to get the trace Z(β). More generally, the lattice approximation of expectation values reads d n wa(w) e S E(w) n A =, where d n w = dw dn we S j, (9.) E(w) and S E (w) = S E (w,...,w n ) is the Euclidean lattice-action entering the discrete path integral (6.2). One could use standard numerical integration routines and approximate an integral by a sum with a finite number of terms M d n wf(w) f(w µ ) w µ. µ= Note that w µ denotes a configuration labelled by µ whereas w j denotes the j th component of the configuration w = (w,...,w n ). the If the space of integration is high-dimensional, it is however advantageous to choose the points w µ at random instead of using a regular set of points. Note, however, that the integrand exp( S) will vary over many orders of magnitude for different configurations w µ (we used that already in the semiclassical approximation). The Monte Carlo method introduced by METROPOLIS, ROSENBLUTH, ROSENBLUTH and TELLER is based on the idea of important sampling [29]. Instead of selecting the points w µ at random they are taken from that region where the dominant contributions to the integrals come from. More precisely, we choose the points according to the Boltzmann distribution P(w) = Z e S E(w), Z = e SE(w) d n w (9.2) 87

CHAPTER 9. SIMULATIONS 9.. Markov Processes and Stochastic Matrices 88 such that the Monte Carlo estimate average Ā for the average of A simply reduces to an arithmetic A Ā = M M A(w µ ), (9.3) µ= wherem is the total number of points (states) generated in the Monte Carlo simulation. 9. Markov Processes and Stochastic Matrices We now discuss a realization of important sampling. It is not quite straightforward, since P is unknown. But it is possible to construct a Markov process generating the M configurations w µ according to given transition probabilities W(w µ w ν ) (the probability that the system, currently in configuration w µ makes a one-step transition into the configuration w ν ). Being a transition probability, the stochastic matrixw must be positive and normalized. To proceed we consider a system with a finite number f of degrees of freedom. We denote the states by s,2,...,f and the transition probabilities by W(s s ) W(s,s ). They are positive, and normalized W(s,s ) 0 (9.4) s W(s,s ) =. (9.5) A f-dimensional matrix W with these properties is called stochastic matrix. In a two-step process from s to s the system must pass through some intermediate state s such that the probability of a transition from s tos in two steps is given by Similarly for an n-step process, we have W n (s,s ) = W (2) (s,s ) = s W(s,s )W(s,s ). (9.6) s s n W(s,s )W(s,s 2 ) W(s n,s ) = s W n (s,s )W(s,s ). (9.7) Now one tries to construct a Markov process such that in the long-time limit n the configurations are distributed according to (9.2). The long-time behavior of the system is determined by the limit ofw n as n. The set of stochastic matrices form a semigroup and every stochastic matrix transforms a stochastic vector (a f-dimensional vector p with non-negative entries P s obeying P s = ) into a stochastic vector. The entry P s ofpis the probability to find the system in the states.

CHAPTER 9. SIMULATIONS 9.. Markov Processes and Stochastic Matrices 89 Consider for example a system with2states and stochastic matrix W = ( ) a a 0 ( a with W n n a n ) = 0 ( ) 0 0 (9.8) for a <. The stochastic matrices W n converges exponentially fast as n. Also note that this matrix has the eigenvalue. Later we shall see that under certain assumptions on W the W n always converge to a stochastic matrix with identical rows. A second and less trivial example is the stochastic matrix for a system with3states, the powers of which are a ( a) ( a) 2 2 W = 0 0 (9.9) 0 0 a n W n 2 ( an ) 2 ( an ) = 0 0 for even n 0 0 and a n W n 2 ( an ) 2 ( an ) = 0 0 for odd n, 0 0 and which again possesses the eigenvalue. Fora < a stochastic vectorpis mapped into p pw n ( 0,p 2 + p 2,p 3 + p ) 2 or ( 0,p 3 + p 2,p 2 + p ) 2 asn, depending on whethernis even or odd. Note thatpw n approaches a periodic orbit exponentially fast and hence the Markov process does not converge. As we shall see, the reason for this lack of convergence is that the minima in all columns ofw are zero. A stochastic matrix has always the eigenvalue. The corresponding right-eigenvector is (,,...,) T. To find the left-eigenvector we consider the sequence p n = n n j=0 pw j. (9.0) Since the set of stochastic vectors is compact this series must have a convergent subsequence n k n k 0 pw j P.

CHAPTER 9. SIMULATIONS 9.. Markov Processes and Stochastic Matrices 90 We multiply withw from right and find Subtracting the two series yields n k pw j PW. n k n k (p pw n k ) P PW Forn k the left hand side tends to zero and hence PW = P. (9.) and therefore any stochastic matrix W possesses at least one fixpoint P that is a eigenvector with eigenvalue. Let us now assume that W possesses at least one column whose entries are bigger or equal to a positive number δ which means that all states have a non-vanishing probability for a transition in a certain state. Such W are called attractive stochastic matrices. The W in the first example above is attractive whereas the second one is not. For an attractive W all statesshave a non-vanishing probability to end up in a given states, that isw(s,s ) > 0 for alls. Now we prove that an attractivew is contractive on vectors of length2. First we note that for two real numberspand p we have such that for two stochastic vectors p p = p+p 2min(p,p ), p p = 2 2 s min(p s,p s ). (9.2) Now we prove that an attractivew is contractive on vectors = (,..., f ) with s = 2 and s = 0. (9.3) First we prove this statement for the difference of two cartesian basis vectors e s, s =,...,f. For that we apply the identity (9.2) to the stochastic vectors e s W and e s W, that is to the rows s and s ofw. For an attractivew we find fors s e s W e s W = 2 2 s min{w(s,s ),W(s,s )} 2 2δ = ( δ) e s e s }{{} =2 with 0 < δ <, (9.4) and this proves thatw is contractive of the difference vectors e s e s. We used min s {W(s,s )W(s,s )} min{w(s,s )(W(s,s )} δ

CHAPTER 9. SIMULATIONS 9.. Markov Processes and Stochastic Matrices 9 wheres belongs to the particular column ofw the elements of which are bigger or equal to δ. Now we prove the contraction property for all vectors in (9.3). Since s: s 0 s: s 0 s s: s<0 s + s: s<0 s = = 2 s = 0 we conclude that s 0 s = and s<0 s =. (9.5) To simplify our notation we denote the non-negative elements of by s and the negative elements by s. Note that the index sets {s} and {s } have no common elements. Because of (9.5) we have = 2 = 2 s s = s s e s e s, (9.6) }{{} =2 where we assumeds s. To bound the norm of W we use s e s = s s e s, s e s = + s s e s, where we made use of (9.5), and this leads to the inequality W = s e s W + s e s W = s s (e s e s )W s s (e s e s )W s s e s e s ( δ), (9.7) where we used the inequality (9.4). Comparing with (9.6) leads to the inequality W ( δ) (9.8) which shows that W is contractive on vectors of the form (9.3). Since this inequality is linear we may drop the condition = 2. Hence W is contractive on all vectors the elements of which add up to zero and in particular on differences of two stochastic vectors. Iterating the inequality we obtain W n ( δ) n. (9.9) Now we apply this inequality to p P, where P is the fixpoint in (9.) and p is an arbitrary stochastic vector. Since the elements ofp P add up to zero we conclude (p P)W n = pw n P n 0

CHAPTER 9. SIMULATIONS 9.2. Detailed Balance, Metropolis Algorithm 92 or equivalently that pw n n P. (9.20) For the stochastic vectors e s the left hand side is just the row s of the limit limw n = W eq such that W eq has identical rows. This means that all elements in a column of W eq are equal, similarly as in the example above. W eq (s,s ) = lim n W n (s,s ) = P s, (9.2) where P s is the element s of P. It follows that both the limit distribution P and matrix W eq are unique. Else there would exist a second fixpoint P of the Markov process with P s = s P sw(s,s ) = lim n s P sw n (s,s ) = s P sp s = P s, and this shows that P is indeed the unique fixpoint. The generalization to systems with continuous degrees of freedom is clear. For the discretized path integral the states of a Markov process are the broken line pathsw = (w,...,w n ) on a time lattice with n sites. The probability P s becomes a probability density P(w). Sums over the discrete index s become integrals over the continuous variables w,...,w n. For the discretized path integral the conditions (9.4) and (9.5) read W(w,w ) 0 and Dw W(w,w ) =. (9.22) The fixpoint condition for the equilibrium distributionp(w) takes the form P(w ) = DwP(w)W(w,w ). (9.23) In Euclidean Quantum Mechanics or in Quantum Statistics the equilibrium distribution P is the the Boltzmann distribution, see (9.2). 9.2 Detailed Balance, Metropolis Algorithm The interesting question is whether we can find a simple stochastic process which has the Boltzmann distribution as its fixpoint. All processes used in actual simulations are constructed with stochastic matrices fulfilling the detailed balance condition P s W(s,s ) = P s W(s,s), (9.24) which means that the transition from s to s is more probable as the inverse process if the equilibrium density is bigger at s than it is at s. If the detailed balance condition holds then P is indeed a fix point ofw since P s W(s,s ) = P s W(s,s) = P s. (9.25) s s

CHAPTER 9. SIMULATIONS 9.2. Detailed Balance, Metropolis Algorithm 93 The detailed balance condition does not fix the stochastic process uniquely. Considerations of computational efficiency dictates its specific form. The local Metropolis- and heatbath algorithm are the most popular once since they can be applied to almost all statistical systems. More recently nonlocal cluster algorithms are also used in cases where local algorithms fail due to critical slowing down. Unfortunately cluster algorithms are not known for lattice gauge theories. In the METROPOLIS algorithm one starts with a certain initial configuration. The better the initial configuration is chosen the less computer time is needed. For example, for a statistical lattice model at high temperature one should choose the variables at different lattice sites (having the lattice approximation (6.2) in mind) randomly, since there is almost no correlation between variables at neighboring sites (see the high temperature expansion). If we are simulating a quantum mechanical system with a deep double well potential then (at low or zero temperature) one better chooses a initial configuration which resembles an instanton. After preparing the initial configuration one takes the first lattice variablew and changes it or leaves it according to the following rules: First one replaces w tentatively by a randomly chosen trial-variablew. IfS decreases (P increases) then the variable at siteis set to the new valuew. If the action increases then a random number r with uniform distribution between 0 and is generated and the variable is changed to w only if exp( S) > r. Otherwise the lattice variable retains its previous valuew. The other sites of the lattice are then probed in exactly the same way. When one arrives at the last lattice site then one has completed one Monte-Carlo sweep. To test whether one is already near equilibrium one measures some expectation values as a function of iteration time. After the measured values have settled (that is we are not too far from the equilibrium distribution) then we may start measuring observables according to (9.3). The Heat-bath algorithm is very similar as the Metropolis algorithm and I refer you to the literature on Monte Carlo simulations for learning more about the various techniques to simulate quantum- or spin system. For an introduction into Monte Carlo (MC) methods, numerical studies and further references I refer you to the nice paper of Creutz and Freedman [30], the text book of NEWMAN and BARKENNA [3] or the text book of BERG [32]. 9.2. Three-state system at finite temperature Let us finally consider a physical system with 3 states. H n = E n n, n =,2,3 with E < E 2 < E 3. (9.26)

CHAPTER 9. SIMULATIONS 9.2. Detailed Balance, Metropolis Algorithm 94 Since the energy decreases for the transitions 2, 3 and 3 2 the corresponding transition amplitudes are constant in the Metropolis algorithm. The inverse transitions cost energy and the probabilities are proportional to the Boltzmann factors b pq = e β(ep Eq), p > q. (9.27) We see that the Markov process has the stochastic matrix W = 2 b 2 b 3 b 2 b 3 b 32 b 32, (9.28) 2 0 and its powers converge to W = e βe e βe 2 e βe 3 e βe e βe 2 e βe 3, (9.29) Z e βe e βe 2 e βe 3 where Z is the partition function, Z = exp( βe )+exp( βe 2 )+exp( βe 3 ). Every initial probability distribution converges to the Boltzmann distribution, P = ( ) e βe,e βe 2,e βe 3. (9.30) Z The following figure shows the convergence to equilibrium for different initial distributions. I used the differences E 2 E = 0.5 and E 3 E 2 = 0.3. kalter Start p p 2 p 3 W p p 2 p 3 W p p 2 p 3 W p p 2 p 3 warmer Start For a cold start with ground state as initial state and a hot start with equal probabilities for the different states we find good convergence to equilibrium. If we start with the most excited state as initial state the the convergence is bad.