II. Probability. II.A General Definitions

Similar documents
Chapter 2. Random Variable. Define single random variables in terms of their PDF and CDF, and calculate moments such as the mean and variance.

Chapter 3 Single Random Variables and Probability Distributions (Part 1)

ΔP(x) Δx. f "Discrete Variable x" (x) dp(x) dx. (x) f "Continuous Variable x" Module 3 Statistics. I. Basic Statistics. A. Statistics and Physics

Random variable X is a mapping that maps each outcome s in the sample space to a unique real number x, x. X s. Real Line

Data, Estimation and Inference

Random variable X is a mapping that maps each outcome s in the sample space to a unique real number x, < x <. ( ) X s. Real Line

Math 180A. Lecture 16 Friday May 7 th. Expectation. Recall the three main probability density functions so far (1) Uniform (2) Exponential.

Probability Distributions - Lecture 5

Chapter 2: The Random Variable

MODULE 6 LECTURE NOTES 1 REVIEW OF PROBABILITY THEORY. Most water resources decision problems face the risk of uncertainty mainly because of the

Stochastic Processes. Review of Elementary Probability Lecture I. Hamid R. Rabiee Ali Jalali

Chapter 1 Statistical Reasoning Why statistics? Section 1.1 Basics of Probability Theory

Scientific Measurement

RS Chapter 2 Random Variables 9/28/2017. Chapter 2. Random Variables

n px p x (1 p) n x. p x n(n 1)... (n x + 1) x!

Introduction to Probability Theory for Graduate Economics Fall 2008

Review of Elementary Probability Lecture I Hamid R. Rabiee

Chapter 3: Random Variables 1

Chapter 1. Sets and probability. 1.3 Probability space

Kinetic Theory 1 / Probabilities

PROBABILITY THEORY. Prof. S. J. Soni. Assistant Professor Computer Engg. Department SPCE, Visnagar

2. A Basic Statistical Toolbox

Basic Concepts and Tools in Statistical Physics

System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models

A Probability Primer. A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes.

Random Variable. Discrete Random Variable. Continuous Random Variable. Discrete Random Variable. Discrete Probability Distribution

A random variable is a variable whose value is determined by the outcome of some chance experiment. In general, each outcome of an experiment can be

Lecture 2. Binomial and Poisson Probability Distributions

2.1 Lecture 5: Probability spaces, Interpretation of probabilities, Random variables

4th IIA-Penn State Astrostatistics School July, 2013 Vainu Bappu Observatory, Kavalur

Lecture 1: Probability Fundamentals

Chapter 3: Random Variables 1

Appendix A PROBABILITY AND RANDOM SIGNALS. A.1 Probability

The random variable 1

Probabilities and distributions

Lecture 2 Binomial and Poisson Probability Distributions

Week 4: Chap. 3 Statistics of Radioactivity

Probability Theory for Machine Learning. Chris Cremer September 2015

1 A brief primer on probability distributions

Relationship between probability set function and random variable - 2 -

Lecture Notes 2 Random Variables. Discrete Random Variables: Probability mass function (pmf)

CSCI-6971 Lecture Notes: Probability theory

3rd IIA-Penn State Astrostatistics School July, 2010 Vainu Bappu Observatory, Kavalur

UNIT 4 MATHEMATICAL METHODS SAMPLE REFERENCE MATERIALS

MATHEMATICS AND STATISTICS

Lecture 16. Lectures 1-15 Review

1 INFO Sep 05

Brownian motion and the Central Limit Theorem

Plotting data is one method for selecting a probability distribution. The following

Computer Science, Informatik 4 Communication and Distributed Systems. Simulation. Discrete-Event System Simulation. Dr.

Kinetic Theory 1 / Probabilities

Statistics for Managers Using Microsoft Excel (3 rd Edition)

Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com

Chapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University

Probability Distribution

Statistics and data analyses

02 Background Minimum background on probability. Random process

1: PROBABILITY REVIEW

If the objects are replaced there are n choices each time yielding n r ways. n C r and in the textbook by g(n, r).

2. AXIOMATIC PROBABILITY

Axioms of Probability

11.4. Differentiating ProductsandQuotients. Introduction. Prerequisites. Learning Outcomes

Statistics, Data Analysis, and Simulation SS 2013

Lecture Notes 2 Random Variables. Random Variable

Notes 12 Autumn 2005

Numerical Methods for Data Analysis

1 Random Variable: Topics

Homework 3 solution (100points) Due in class, 9/ (10) 1.19 (page 31)

Statistics for Managers Using Microsoft Excel/SPSS Chapter 4 Basic Probability And Discrete Probability Distributions

Statistics Random Variables

Advanced Herd Management Probabilities and distributions

Answers to All Exercises

EE/CpE 345. Modeling and Simulation. Fall Class 5 September 30, 2002

Lecture 2: Repetition of probability theory and statistics

2.7 The Gaussian Probability Density Function Forms of the Gaussian pdf for Real Variates

g(y)dy = f(x)dx. y B If we write x = x(y) in the second integral, then the change of variable technique gives g(y)dy = f{x(y)} dx dy dy.

STAT:5100 (22S:193) Statistical Inference I

Physics 403 Probability Distributions II: More Properties of PDFs and PMFs

04. Random Variables: Concepts

Statistics, Data Analysis, and Simulation SS 2017

Probability Distribution

Error analysis in biology

Math 416 Lecture 3. The average or mean or expected value of x 1, x 2, x 3,..., x n is

Random Variables. P(x) = P[X(e)] = P(e). (1)

Frank Porter April 3, 2017

Introduction. Probability and distributions

Statistics, Probability Distributions & Error Propagation. James R. Graham

CS 229: Lecture 7 Notes

Probability Theory and Applications

213 Midterm coming up

Probability (10A) Young Won Lim 6/12/17

Chapter 4 Multiple Random Variables

Chapter 3. Chapter 3 sections

ELEG 3143 Probability & Stochastic Process Ch. 2 Discrete Random Variables

EXAM. Exam #1. Math 3342 Summer II, July 21, 2000 ANSWERS

CHAPTER 3 RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS. 3.1 Concept of a Random Variable. 3.2 Discrete Probability Distributions

PHYS 6710: Nuclear and Particle Physics II

2.3 Estimating PDFs and PDF Parameters

LECTURE NOTES FYS 4550/FYS EXPERIMENTAL HIGH ENERGY PHYSICS AUTUMN 2013 PART I A. STRANDLIE GJØVIK UNIVERSITY COLLEGE AND UNIVERSITY OF OSLO

Some Statistics. V. Lindberg. May 16, 2007

Transcription:

II. Probability II.A General Definitions The laws of thermodynamics are based on observations of macroscopic bodies, and encapsulate their thermal properties. On the other hand, matter is composed of atoms and molecules whose motions are governed by fundamental laws of classical or quantum mechanics. It should be possible, in principle, to derive the behavior of a macroscopic body from the knowledge of its components. This is the problem addressed by kinetic theory in the following lectures. Actually, describing the full dynamics of the enormous number of particles involved is quite a daunting task. As we shall demonstrate, for discussing equilibrium properties of a macroscopic system, full knowledge of the behavior of its constituent particles is not necessary. All that is required is the likelihood that the particles are in a particular microscopic state. Statistical mechanics is thus an inherently probabilities description of the system. Familiarity with manipulations of probabilities is therefore an important prerequisite to statistical mechanics. Our purpose here is to review some important results in the theory of probability, and to introduce the notations that will be used in the rest of the course. The entity under investigation is a random variable, which has a set of possible outcomes S { 1, 2, }. The outcomes may be discrete as in the case of a coin toss, S coin = {head, tail}, or a dice throw, S dice = {1, 2, 3, 4, 5, 6}, or continuous as for the velocity of a particle in a gas, S v = { < v, v y, v z < }, or the energy of an electron in a metal at zero temperature, S ǫ = {0 ǫ ǫ F }. An event is any subset of outcomes E S, and is assigned a probability pe), e.g. p dice {1}) = 1/6, or p dice {1, 3}) = 1/3. From an aiomatic point of view, the probabilities must satisfy the following conditions: i) Positivity: pe) 0, i.e. all probabilities must be non-zero. ii) Additivity: pa or B) = pa) + pb), if A and B are disconnected events. iii) Normalization: ps) = 1, i.e. the random variable must have some outcome in S. From a practical point of view, we would like to know how to assign probability values to various outcomes. There are two possible approaches: 1) Objective probabilities are obtained eperimentally from the relative frequency of the occurrence of an outcome in many tests of the random variable. If the random process is repeated N times, and the event A occurs N A times, then N A pa) = lim N N. 25

For eample, a series of N = 100, 200, 300 throws of a dice may result in N 1 = 19, 30, 48 occurrences of 1. The ratios.19,.15,.16 provide an increasingly more reliable estimate of the probability p dice {1}). 2) Subjective probabilities provide a theoretical estimate based on the uncertainties related to lack of precise knowledge of outcomes. For eample, the assessment p dice {1}) = 1/6, is based on the knowledge that there are si possible outcomes to a dice throw, and that in the absence of any prior reason to believe that the dice is biased, all si are equally likely. All assignments of probability in Statistical Mechanics are subjectively based.the consequences of such subjective assignments of probability have to be checked against measurements, and they may need to be modified as more information about the outcomes becomes available. II.B One Random Variable As the properties of a discrete random variable are rather well known, here we focus on continuous random variables, which are more relevant to our purposes. Consider a random variable, whose outcomes are real numbers, i.e. S = { < < }. The cumulative probability function CPF) P), is the probability of an outcome with any value less than, i.e. P) = prob.e, ). P) must be a monotonically increasing function of, with P) = 0 and P+ ) = 1. The probability density function PDF) is defined by p) dp)/d. Hence, p)d = prob.e, + d). As a probability density, it is positive, and normalized such that prob.s) = d p) = 1. II.1) Note that since p) is a probability density, it has no upper bound, i.e. 0 < p) <. The epectation value of any function F), of the random variable is F) = d p)f). II.2) The function F) is itself a random variable, with an associated PDF of p F f)df = prob.f) f, f + df). There may be multiple solutions i, to the equation F) = f, and p F f)df = p i )d i, = p F f) = p i ) d df i i =i. II.3) 26

The factors of d/df are the Jacobians associated with the change of variables from to F. For eample, consider p) = λ ep λ )/2, and the function F) = 2. There are two solutions to F) = f, located at ± ± f 1/2 /2. Hence, P F f) = λ λ 2 ep ) f 1 2 f = ± f, with corresponding Jacobians ) + 1 2 f = λ ep λ f ) 2, f for f > 0, and p F f) = 0 for f < 0. Note that p F f) has an integrable) divergence at f = 0. Moments of the PDF are epectation values for powers of the random variable. The n th moment is m n n = dp) n. II.4) The characteristic function, is the generator of moments of the distribution. It is simply the Fourier transform of the PDF, defined by pk) = e ik = dp) e ik. II.5) The PDF can be recovered from the characteristic function through the inverse Fourier transform p) = 1 2π dk pk) e +ik. Moments of the distribution are obtained by epanding pk) in powers of k, II.6) ik) n pk) = n = n=0 n=0 n. II.7) Moments of the PDF around any point 0 can also be generated by epanding e ik0 pk) = e ik 0) = n=0 0 ) n. II.8) The cumulant generating function is the logarithm of the characteristic function. Its epansion generates the cumulants of the distribution defined through ln pk) = n c. II.9) 27

Relations between moments and cumulants can be obtained by epanding the logarithm of pk) in eq.ii.7), and using ln1 + ǫ) = 1) n+1 ǫn n. II.10) The first four cumulants are called the mean, variance, skewness, and curtosis of the distribution respectively, and are obtained from the moments as c =, 2 c = 2 2, 3 c = 3 3 2 + 2 3, 4 c = 4 4 3 3 2 2 + 12 2 2 6 4. The cumulants provide a useful and compact way of describing a PDF. II.11) An important theorem allows easy computation of moments in terms of the cumulants: Represent the n th cumulant graphically as a connected cluster of n points. The m th moment is then obtained by summing all possible subdivisions of m points into groupings of smaller connected or disconnected) clusters. The contribution of each subdivision to the sum is the product of the connected cumulants that it represents. Using this result the first four moments are easily computed as = c, 2 = 2 + c 2 c, 3 = 3 + 3 2 c c c + 3 c, 4 = 4 c + 4 3 c c + 3 2 2 c + 6 2 c 2 c + 4 c. II.12) This theorem, which is the starting point for various diagrammatic computations is statistical mechanics and field theory, is easily proved by equating the epression in eqs. II.7) and II.9) for pk) ik) m m ik) n = ep n m! c = n m=0 ik) np n p n p n! n ) pn c. II.13) Equating the powers of ik) m on the two sides of the above epression leads to m = {p n } n m! p n!) p n n p n c. II.14) The sum is restricted such that np n = m, and leads to the graphical interpretation given above, as the numerical factor is simply the number of ways of breaking m points into {p n } clusters of n points. 28

II.C Some Important Probability Distributions The properties of three commonly encountered probability distributions are eamined in this section. 1) The normal Gaussian) distribution describes a continuous real random variable, with p) = 1 ep λ)2 2πσ 2 2σ 2 The corresponding characteristic function also has a Gaussian form, pk) = d 1 ep λ)2 2πσ 2 2σ 2 ik = ep ikλ k2 σ 2 2. II.15). II.16) Cumulants of the distribution can be identified from ln pk) = ikλ k 2 σ 2 /2, using eq.ii.9), as c = λ, 2 c = σ2, 3 c = 4 c = = 0. II.17) The normal distribution is thus completely specified by its two first cumulants. This makes the computation of moments using the cluster epansion eqs.ii.12)) particularly simple, and =λ, 2 =σ 2 + λ 2, 3 =3σ 2 λ + λ 3, 4 =3σ 4 + 6σ 2 λ 2 + λ 4,. II.18) The normal distribution serves as the starting point for most perturbative computations in field theory. The vanishing of higher cumulants implies that all graphical computations involve only products of one point, and two point known as propagators) clusters. 2) The binomial distribution: Consider a random variable with two outcomes A and B e.g. a coin toss) of relative probabilities p A and p B = 1 p A. The probability that in N trials the event A occurs eactly N A times e.g. 5 heads in 12 coin tosses), is given by the binomial distribution N p N N A ) = N A ) p N A A pn N A B. II.19) The prefactor, N N A ) = N! N A!N N A )!, 29 II.20)

is just the coefficient obtained in the binomial epansion of p A +p B ) N, and gives the number of possible orderings of N A events A and N B = N N A events B. The characteristic function for this discrete distribution is given by p N k) = e ikn A = N N A =0 The resulting cumulant generating function is N! N A!N N A )! pn A A pn N A B e ikn A = p A e ik + p B ) N. II.21) ln p N k) = N ln p A e ik + p B ) = N ln p1 k), II.22) where ln p 1 k) is the cumulant generating function for a single trial. Hence, the cumulants after N trials are simply N times the cumulants in a single trial. In each trial, the allowed values of N A are 0 and 1 with respective probabilities p B and p A, leading to N l A = pa, for all l. After N trials the first two cumulants are N A c = Np A, N 2 A c = Np A p 2 A) = Np A p B. II.23) A measure of fluctuations around the mean is provided by the standard deviation, which is the square root of the variance. While the mean of the binomial distribution scales as N, its standard deviation only grows as N. Hence, the relative uncertainty becomes smaller for large N. The binomial distribution is straightforwardly generalized to a multinomial distribution, when the several outcomes {A, B,, M} occur with probabilities {p A, p B,, p M }. The probability of finding outcomes {N A, N B,, N M } in a total of N = N A +N B +N M trials is p N {N A, N B,, N M }) = N! N A!N B! N M! pn A A pn B B pn M M. II.24) 3) The Poisson distribution: The classical eample of a Poisson process is radioactive decay. Observing a piece of radioactive material over a time interval T shows that: a) The probability of one and only one event decay) in the interval t, t + dt is proportional to dt as dt 0, b) The probabilities of events at different intervals are independent of each other. The probability of observing eactly M decays in the interval T is given by the Poisson distribution. It is obtained as a limit of the binomial distribution by subdividing the interval into N = T/dt 1 segments of size dt. In each segment, an event occurs with probability p = αdt, and there is no event with probability q = 1 αdt. As the probability 30

of more than one event in dt is too small to consider, the process is equivalent to a binomial one. Using eq.ii.21), the characteristic function is given by pk) = pe ik + q ) n = lim dt 0 1 + αdt e ik 1 ) T/dt = ep αe ik 1)T. II.25) The Poisson PDF is obtained from the inverse Fourier transform in eq.ii.6) as p) = dk 2π ep α e ik 1 ) T + ik = e αt dk 2π eik using the power series for the eponential. The integral over k is leading to p αt ) = dk 2π eik M) = δ M), m=0 αt αt)m e δ M). M! M=0 αt) M e ikm, M! II.26) II.27) II.28) The probability of M events is thus p αt M) = e αt αt) M /M!. The cumulants of the distribution are obtained from the epansion ln p αt k) = αte ik 1) = αt, = M n c = αt. II.29) All cumulants have the same value, and the moments are obtained from eqs.ii.12) as M = αt), M 2 = αt) 2 + αt), M 3 = αt) 3 + 3αT) 2 + αt). II.30) Eample: Assuming that stars are randomly distributed in the galay clearly unjustified) with a density n, what is the probability that the nearest star is at a distance R? Since, the probability of finding a star in a small volume dv is ndv, and they are assumed to be independent, the number of stars in a volume V is described by a Poisson process as in eq.ii.28), with α = n. The probability pr), of encountering the first star at a distance R is the product of the probabilities p nv 0), of finding zero stars in the volume V = 4πR 3 /3 around the origin, and p ndv 1), of finding one star in the shell of volume dv = 4πR 2 dr at a distance R. Both p nv 0) and p ndv 1) can be calculated from eq.ii.28), and pr)dr = p nv 0) p ndv 1) =e 4πR3n/3 e 4πR2ndR 4πR 2 ndr, = pr) =4πR 2 n ep 4π ) 3 R3 n. II.31) 31