An Outline of (Classical) Statistical Mechanics and Related Concepts in Machine Learning

Similar documents
(# = %(& )(* +,(- Closed system, well-defined energy (or e.g. E± E/2): Microcanonical ensemble

9.1 System in contact with a heat reservoir

MASSACHUSETTS INSTITUTE OF TECHNOLOGY Physics Department Statistical Physics I Spring Term 2013 Notes on the Microcanonical Ensemble

2m + U( q i), (IV.26) i=1

to satisfy the large number approximations, W W sys can be small.

Grand Canonical Formalism

IV. Classical Statistical Mechanics

Nanoscale simulation lectures Statistical Mechanics

1. Thermodynamics 1.1. A macroscopic view of matter

1 The fundamental equation of equilibrium statistical mechanics. 3 General overview on the method of ensembles 10

Thermodynamics of the nucleus

G : Statistical Mechanics

PHY 5524: Statistical Mechanics, Spring February 11 th, 2013 Midterm Exam # 1

In-class exercises. Day 1

Phase transitions for particles in R 3

UNIVERSITY OF OSLO FACULTY OF MATHEMATICS AND NATURAL SCIENCES

summary of statistical physics

4.1 Constant (T, V, n) Experiments: The Helmholtz Free Energy

Introduction Statistical Thermodynamics. Monday, January 6, 14

ELEMENTARY COURSE IN STATISTICAL MECHANICS

Chapter 4: Going from microcanonical to canonical ensemble, from energy to temperature.

ChE 210B: Advanced Topics in Equilibrium Statistical Mechanics

Contents. 1 Introduction and guide for this text 1. 2 Equilibrium and entropy 6. 3 Energy and how the microscopic world works 21

Elements of Statistical Mechanics

PHYS 352 Homework 2 Solutions

Part II: Statistical Physics

1 Foundations of statistical physics

(i) T, p, N Gibbs free energy G (ii) T, p, µ no thermodynamic potential, since T, p, µ are not independent of each other (iii) S, p, N Enthalpy H

Part II: Statistical Physics

ChE 503 A. Z. Panagiotopoulos 1

III. Kinetic Theory of Gases

A Brief Introduction to Statistical Mechanics

Removing the mystery of entropy and thermodynamics. Part 3

i=1 n i, the canonical probabilities of the micro-states [ βǫ i=1 e βǫn 1 n 1 =0 +Nk B T Nǫ 1 + e ǫ/(k BT), (IV.75) E = F + TS =

Classical Statistical Mechanics: Part 1

Introduction to the Renormalization Group

Lecture 4: Entropy. Chapter I. Basic Principles of Stat Mechanics. A.G. Petukhov, PHYS 743. September 7, 2017

arxiv:cond-mat/ v1 10 Aug 2002

Derivation of the GENERIC form of nonequilibrium thermodynamics from a statistical optimization principle

The Kelvin-Planck statement of the second law

Statistical Mechanics in a Nutshell

Statistical Thermodynamics and Monte-Carlo Evgenii B. Rudnyi and Jan G. Korvink IMTEK Albert Ludwig University Freiburg, Germany

Part II Statistical Physics

6.730 Physics for Solid State Applications

Microcanonical Ensemble

Thermodynamics of feedback controlled systems. Francisco J. Cao

Statistical Mechanics. Atomistic view of Materials

Dually Flat Geometries in the State Space of Statistical Models

[S R (U 0 ɛ 1 ) S R (U 0 ɛ 2 ]. (0.1) k B

Statistical Mechanics Notes. Ryan D. Reece

1 Phase Spaces and the Liouville Equation

arxiv: v2 [cond-mat.stat-mech] 8 Feb 2016

Temperature Fluctuations and Entropy Formulas

Statistical Mechanics

Statistical Mechanics I

Thermodynamics, Statistical Mechanics and Entropy

Brief Review of Statistical Mechanics

Statistical mechanics of classical systems

Derivation of Van der Waal s equation of state in microcanonical ensemble formulation

MACROSCOPIC VARIABLES, THERMAL EQUILIBRIUM. Contents AND BOLTZMANN ENTROPY. 1 Macroscopic Variables 3. 2 Local quantities and Hydrodynamics fields 4

Statistical Mechanics

Theoretical Statistical Physics

Statistical Mechanics and Information Theory

Summer Lecture Notes Thermodynamics: Fundamental Relation, Parameters, and Maxwell Relations

Non-equilibrium phenomena and fluctuation relations

Concepts in Materials Science I. StatMech Basics. VBS/MRC Stat Mech Basics 0

The fine-grained Gibbs entropy

First Law Limitations

Basics of Statistical Mechanics

2. Thermodynamics. Introduction. Understanding Molecular Simulation

Javier Junquera. Statistical mechanics

Einstein s early work on Statistical Mechanics

An Entropy depending only on the Probability (or the Density Matrix)

Principles of Statistical Mechanics

Emergent Fluctuation Theorem for Pure Quantum States

Lecture 1: Historical Overview, Statistical Paradigm, Classical Mechanics

Verschränkung versus Stosszahlansatz: The second law rests on low correlation levels in our cosmic neighborhood

Property of Tsallis entropy and principle of entropy increase

Thus, the volume element remains the same as required. With this transformation, the amiltonian becomes = p i m i + U(r 1 ; :::; r N ) = and the canon

Molecular Modeling of Matter

Introduction to Statistical Thermodynamics

Chapter 2 Ensemble Theory in Statistical Physics: Free Energy Potential

PHAS2228 : Statistical Thermodynamics

Phys Midterm. March 17

Entropy Gradient Maximization for Systems with Discoupled Time

Table of Contents [ttc]

fiziks Institute for NET/JRF, GATE, IIT-JAM, JEST, TIFR and GRE in PHYSICAL SCIENCES

Physics 172H Modern Mechanics

Assignment 3. Tyler Shendruk February 26, 2010

CH 240 Chemical Engineering Thermodynamics Spring 2007

although Boltzmann used W instead of Ω for the number of available states.

First Problem Set for Physics 847 (Statistical Physics II)

Statistical Physics. The Second Law. Most macroscopic processes are irreversible in everyday life.

Thermal pure quantum state

Thermodynamics and Statistical Physics WS 2018/19

TSTC Lectures: Theoretical & Computational Chemistry

Statistical Mechanics of Granular Systems

Chapter 10: Statistical Thermodynamics

Hydrodynamics. Stefan Flörchinger (Heidelberg) Heidelberg, 3 May 2010

U = γkt N. is indeed an exact one form. Thus there exists a function U(S, V, N) so that. du = T ds pdv. U S = T, U V = p

Transcription:

An Outline of (Classical) Statistical Mechanics and Related Concepts in Machine Learning Chang Liu Tsinghua University June 1, 2016 1 / 22

What is covered What is Statistical mechanics developed for? What is its basic principle? What does it achieve? What tools/techniques is employed? What is the origin of some concepts in machine learning? Entropy, Boltzmann distribution (canonical distribution), (Helmholtz) free energy, etc. 2 / 22

Overview 1 Introduction 2 The Microcanonical Ensemble 3 The Canonical Ensemble 4 Free Energies 3 / 22

Introduction Introduction 4 / 22

Introduction What is statistical mechanics On the study of statistical properties of a macroscopic system in equilibrium that consists of a huge number of particles: Thermodynamics: based on the four thermodynamic laws summarized from experiments. Week dependence on microscopic laws of motion. Studies the Relations of thermodynamic quantities (e.g. P V = Nk B T, E = 3 2 Nk BT ) (Equilibrium) statistical mechanics: based on statistics, conclusions from microscopic laws of motion, and ergodic hypothesis. Studies the ensembles (an ensemble is the distribution over microstates of a specific macroscopic system). Relations of thermodynamic quantities in thermodynamics can be derived by taking the large system limit in statistical mechanics. 5 / 22

Introduction The basic principle of statistical mechanics The basic principle of statistical mechanics: the probability density over microstates of an isolated system in equilibrium is constant. The principle is supported by: A microstate of an isolated (classical) system comprising N particles: the most detailed description of the system at some instant: s = (Q, P ) = (q 1x, q 1y, q 1z, q 2x,..., q Nz, p 1x, p 1y, p 1z,..., p Nz ) is a point in phase-space R 6N. This description arises from classical mechanics. The time evolution of a microstate is governed by Hamilton s equation: q = p H(s), ṗ = q H(s), where H(s) is the Hamiltonian, a function of a microstate s, and usually the total energy of the system. Hamiltonian is conserved while evoluting: dh(s)/dt = q H q + p H ṗ = q H p H p H q H = 0. Liouville s theorem: dρ/dt = 0. 6 / 22

Introduction The basic principle of statistical mechanics Ergodic hypothesis (first proposed by L. E. Boltzmann): informally, from any initial microstate, the time evolution will lead the system to pass every other microstates of the same energy. For some systems like a three-body system, although chaotic, but not ergodic. For some systems like the hard sphere gas, ergodicity can be proven. We believe ergodicity holds for most systems with large numbers of interacting particles. For the equilibrium distribution over microstates, starting with state s(0), for any other microstate A, there is a time t s.t. s(t) = A (ergodicity), and ρ(s(0), 0) = ρ(s(t), t) = ρ(a, t) (Liouville s theorem), ρ(a, t) = ρ(a, 0) (equilibrium), so ρ(a, 0) = ρ(s(0), 0). The equilibrium distribution is constant over microstates for an isolated system. 7 / 22

The Microcanonical Ensemble The Microcanonical Ensemble 8 / 22

The Microcanonical Ensemble The microcanonical ensemble The distribution over microstates of an isolated system (fixed energy, volume and particle number). According to the basic principle, the density ρ(s) δ(e H(s)). Define Ω(E) = δ(e H(s))ds = d de H(s) E ds (the volume of constant-energy region in the phase space / the number of microstates of constant energy), then ρ(s) = 1 Ω(E) δ(e H(s)). 9 / 22

The Microcanonical Ensemble Relation between ρ(s) and ρ(e) From the basic principle, ρ(s) = f(h(s)). For a general function g(e), we have ˆ ˆ ( ˆ ) g(h(s))ds = g(e)δ(e H(s))dE ds ˆ = g(e)δ(e H(s))dsdE = g(e)ω(e)de. (1) Apply the above conclusion: ˆ ˆ ρ(e) = ρ(s, E)ds = f(h(s))δ(e H(s))ds ˆ Eqn. (1) ======== f(y)δ(e y)ω(y)dy = f(e)ω(e). (2) 10 / 22

The Microcanonical Ensemble Introducing temperature and equilibrium entropy Consider an isolated system consisting of two subsystems that share the total energy E (in heat contact). What is the energy of each subsystem when in equilibrium? Let s 1 and s 2 be an microstate of each subsystem. The whole system is a microcanonical ensemble: ρ(s 1, s 2 ) = 1 Ω(E) δ(e H 1(s 1 ) H 2 (s 2 )). So ˆ ρ(s 1 ) = ρ(s 1, s 2 )ds 2 = 1 ˆ δ(e H 1 (s 1 ) H 2 (s 2 ))ds 2 Eqn. (1) ======== 1 ˆ Ω(E) Ω(E) δ(e H 1 (s 1 ) E 2 )Ω 2 (E 2 )de 2 = 1 Ω(E) Ω 2(E H 1 (s 1 )), (3) and from Eqn. (2) ρ(e 1 ) = 1 Ω(E) Ω 1(E 1 )Ω 2 (E E 1 ). The mode of the energy distribution should satisfy: 0 = dρ(e 1) de 1 = 1 Ω(E) ( dω1 (E 1 ) Ω 2 (E 2 ) Ω 1 (E 1 ) dω 2(E 2 ) ) de 1 de 2 11 / 22

The Microcanonical Ensemble Introducing temperature and equilibrium entropy It can be shown that ρ(e 1 ) is very sharp so at equilibrium E 1 stays at the mode: d log Ω 1 (E 1 ) = d log Ω 2(E 2 ). de 1 de 2 If d log Ω 1(E 1 ) de 1 < d log Ω 2(E 2 ) de 2, then dρ(e 1) de 1 < 0, so E 1 tends to decrease. So d log Ω(E) the subsystem with a smaller de tends to give out its energy. We know that the two systems have the same temperature in equilibrium,and a higher temperature indicates the tend to give out energy, so we define the temperature T as 1 d log Ω(E) =. T de Define the equilibrium entropy S(E, V, N) = k B log Ω(E, V, N) and the temperature is formally defined as 1 T = ( ) S. E V,N 12 / 22

The Microcanonical Ensemble An intuitive interpretation of ρ(e 1 ) to be a very sharp distribution Expand ρ(e 1 ) around its mode E 1: ρ(e 1 ) Ω 1 (E 1 )Ω 2 (E E 1 ) = exp ( ) S 1 (E 1 )/k B + S 2 (E E 1 )/k B [( = exp S 1 (E1) + 1 2 S 1 2 E1 2 (E 1 E1) 2 + S 2 (E2) + 1 2 ) ] S 2 2 E2 2 (E 1 E1) 2 /k B exp[ (E 1 E1) 2 /(2σ E )], where σ 2 E = k B/( 2 S 1 / E 2 1 + 2 S 2 / E 2 2). E and S are extensive quantities (scale linearly with system size N), so 2 S/ E 2 1/N and the relative fluctuation σ E /E 1 N/N = 1/ N vanishes in the limit N. 13 / 22

The Microcanonical Ensemble The history of entropy Entropy in thermodynamics: Entropy is first proposed in thermodynamics by Rudolf Clausius. The fact that C δq T = 0 for any reversible loop C indicating that there exists a state function S and the difference for two states A and B is S(B) S(A) = l δq T for any reversible path l from A to B. Entropy in equilibrium statistical mechanics: Namely the above definition S = k B log Ω, which is proposed by Planck based on the idea of Boltzmann. For thermodynamic entropy, ds = δq T. For Boltzmann s entropy, T = ( E S )V,N = δq ds, so ds = δq T also holds. That is the relation between the two definitions. 14 / 22

The Microcanonical Ensemble The history of entropy Shannon s entropy in information theory: S = k S i p i log p i = ρ(x) log ρ(x)dx. Boltzmann s entropy: S = k B log Ω = k B log(1/ω) = k Ω B i=1 (1/Ω) log(1/ω). Shannon s entropy is the only continuous function of p = (p 1,..., p N ) that satisfies: 1)S( 1 N,..., 1 N ) S(p 1,..., p N ), and the equality holds if and only if p is uniform; 2) S(p 1,..., p N 1, 0) = S(p 1,..., p N 1 ); 3) for a joint distribution over (x, y), E y [S(x y)] = S(x, y) S(y). 15 / 22

The Canonical Ensemble The Canonical Ensemble 16 / 22

The Canonical Ensemble The canonical ensemble The distribution over microstates of a system with fixed temperature, volume and particle number. The only theoretical way to implement such a system is to let it in contact with another system large enough to maintain temperature while exchanging energy with the concerned system. The large enough system is called a heat bath (thermal energy reservoir). Deriving the distribution over microstates: denote the concerned system as 1 and the heat bath as 2. Then according to Eqn. (3), ρ(s 1 ) Ω 2 (E H 1 (s 1 )). According to the large heat bath assumption, S 2 / E 2 = 1/T is fixed. So log Ω 2 / E 2 = 1/k B T, solved as Ω 2 (E 2 ) exp(e 2 /k B T ). So ρ(s 1 ) exp( H 1 (s 1 )/k B T ), and ρ(s) = 1 exp( βh(s)), (4) Z(β) where β = 1/k B T and Z(β) is the partition function. This is the Boltzmann distribution, or canonical distribution. 17 / 22

The Canonical Ensemble The canonical ensemble The Boltzmann distribution is in the exponential family with sufficient statistics H(s) and natural parameter β. So E[E] = log Z/ β. Entropy S = ρ(s) log ρ(s)ds = E[E]/T + k B log Z. Define the Helmholtz free energy A = k B T log Z = E[E] T S, then S = A T. Generally thermodynamic properties of a canonical ensemble are determined by the Helmholtz free energy through its derivatives. 18 / 22

Free Energies Free Energies 19 / 22

Free Energies Free energies The free energy of a specific system determines thermodynamic properties of the system. The free energy of a specific system is the part of internal energy that can be converted into useful work. E.g. The Helmholtz free energy: A = E[E] T S is the total internal energy excluding the heat that has to be transmitted into a cold bath when doing work, thus the available work. A free energy F (x) integrates out other degrees of freedom (variables). exp( F (x)/k B T ) is the volume of the region in phase space with constant x. E.g. For the Helmholtz free energy, exp( A(T )/k B T ) = Z(T ) is the volume of constant-temperature region in phase space. It is the part that region with constant given temperature T takes in the phase space. 20 / 22

Free energies Free Energies Free energies are Legendre transforms of energy. The Legendre transform of a convex function f(x) is f (p) = min x f(x) xp. It can be extended to Fenchel transform for convex conjugate. For differentiable f(x), f (p) = (f(x) xp) p=f (x). The Helmholtz free energy A(T ) is the Legendre transform of E(S): A(T ) = (E(S) ST ) since T = E/ S. 21 / 22

Free Energies Thanks! 22 / 22