Lecture 3: Central Limit Theorem

Similar documents
Lecture 3: Central Limit Theorem

The Quantum Harmonic Oscillator

Return probability on a lattice

A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring 2011

Lecture 25: Large Steps and Long Waiting Times

Notes: Most of the material presented in this chapter is taken from Jackson, Chap. 2, 3, and 4, and Di Bartolo, Chap. 2. 2π nx i a. ( ) = G n.

A Detailed Look at a Discrete Randomw Walk with Spatially Dependent Moments and Its Continuum Limit

Corrections to the Central Limit Theorem for Heavy-tailed Probability Densities

Let's transfer our results for conditional probability for events into conditional probabilities for random variables.

Physics 137A Quantum Mechanics Fall 2012 Midterm II - Solutions

1 Characteristic functions

1 Introduction. 2 Diffusion equation and central limit theorem. The content of these notes is also covered by chapter 3 section B of [1].

Performance Evaluation of Generalized Polynomial Chaos

PRINCIPLES OF STATISTICAL INFERENCE

Multivariate Distributions

Superposition of electromagnetic waves

Eigenmodes for coupled harmonic vibrations. Algebraic Method for Harmonic Oscillator.

Optional Problems on the Harmonic Oscillator

Lecture 7. 1 Wavepackets and Uncertainty 1. 2 Wavepacket Shape Changes 4. 3 Time evolution of a free wave packet 6. 1 Φ(k)e ikx dk. (1.

04. Random Variables: Concepts

Lecture 3: From Random Walks to Continuum Diffusion

Recurrences and Full-revivals in Quantum Walks

Applied Probability and Stochastic Processes

arxiv:cond-mat/ v2 [cond-mat.stat-mech] 23 Feb 1998

An operator is a transformation that takes a function as an input and produces another function (usually).

Lecture 6 Quantum Mechanical Systems and Measurements

Simple Harmonic Oscillator

6.1 Moment Generating and Characteristic Functions

Introduction to Mathematical Physics

Massachusetts Institute of Technology Physics Department

Probability, CLT, CLT counterexamples, Bayes. The PDF file of this lecture contains a full reference document on probability and random variables.

8.04 Spring 2013 March 12, 2013 Problem 1. (10 points) The Probability Current

Laplace s Method for Ordinary Differential Equations

Module 3. Function of a Random Variable and its distribution

Linear Algebra Review (Course Notes for Math 308H - Spring 2016)

The Basic Properties of Surface Waves

Estimation of the Success Probability of Random Sampling by the Gram-Charlier Approximation

From Gram-Charlier Series to Orthogonal Polynomials

Chem 3502/4502 Physical Chemistry II (Quantum Mechanics) 3 Credits Spring Semester 2006 Christopher J. Cramer. Lecture 9, February 8, 2006

Theory and Applications of Stochastic Systems Lecture Exponential Martingale for Random Walk

Page 404. Lecture 22: Simple Harmonic Oscillator: Energy Basis Date Given: 2008/11/19 Date Revised: 2008/11/19

Harmonic Oscillator. Robert B. Griffiths Version of 5 December Notation 1. 3 Position and Momentum Representations of Number Eigenstates 2

Inference For High Dimensional M-estimates: Fixed Design Results

Quantum Mechanics for Mathematicians: Energy, Momentum, and the Quantum Free Particle

The Sommerfeld Polynomial Method: Harmonic Oscillator Example

M4L5. Expectation and Moments of Functions of Random Variable

221A Lecture Notes Convergence of Perturbation Theory

1 Introduction. 2 Measure theoretic definitions

Econ 508B: Lecture 5

Generalized Central Limit Theorem and Extreme Value Statistics

Mod-φ convergence I: examples and probabilistic estimates

Mathematics for Engineers and Scientists

Review (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology

Econ 2148, fall 2017 Gaussian process priors, reproducing kernel Hilbert spaces, and Splines

Fourier transforms, Generalised functions and Greens functions

Numerical Analysis for Statisticians

QM and Angular Momentum

Fluctuations from the Semicircle Law Lecture 1

Algorithms for Uncertainty Quantification

Multivariate Normal-Laplace Distribution and Processes

Lecture 2: Repetition of probability theory and statistics

Recurrences in Quantum Walks

Special Functions of Mathematical Physics

Review (Probability & Linear Algebra)

Algorithms for Uncertainty Quantification

PROPERTIES OF THE EMPIRICAL CHARACTERISTIC FUNCTION AND ITS APPLICATION TO TESTING FOR INDEPENDENCE. Noboru Murata

IEOR 3106: Introduction to Operations Research: Stochastic Models. Fall 2011, Professor Whitt. Class Lecture Notes: Thursday, September 15.

Mathematical Methods for Engineers and Scientists 1

Lecture 10 February 25, 2010

MATHEMATICAL ENGINEERING TECHNICAL REPORTS. A Superharmonic Prior for the Autoregressive Process of the Second Order

I. QUANTUM MONTE CARLO METHODS: INTRODUCTION AND BASICS

Mathematical Physics Homework 10

Today: Fundamentals of Monte Carlo

Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science Quantum Optical Communication

From The Picture Book of Quantum Mechanics, S. Brandt and H.D. Dahmen, 4th ed., c 2012 by Springer-Verlag New York.

Quantum Physics III (8.06) Spring 2007 FINAL EXAMINATION Monday May 21, 9:00 am You have 3 hours.

Physics 115A: Statistical Physics

Collocation based high dimensional model representation for stochastic partial differential equations

Long-range dependence

Lecture 2. Spring Quarter Statistical Optics. Lecture 2. Characteristic Functions. Transformation of RVs. Sums of RVs

1 Distributions (due January 22, 2009)

MATH3383. Quantum Mechanics. Appendix D: Hermite Equation; Orthogonal Polynomials

SOLUTIONS for Homework #2. 1. The given wave function can be normalized to the total probability equal to 1, ψ(x) = Ne λ x.

EECS 598: Statistical Learning Theory, Winter 2014 Topic 11. Kernels

Probability Theory I: Syllabus and Exercise

EEL 5544 Noise in Linear Systems Lecture 30. X (s) = E [ e sx] f X (x)e sx dx. Moments can be found from the Laplace transform as

3. Probability and Statistics

The circular law. Lewis Memorial Lecture / DIMACS minicourse March 19, Terence Tao (UCLA)

A new class of sum rules for products of Bessel functions. G. Bevilacqua, 1, a) V. Biancalana, 1 Y. Dancheva, 1 L. Moi, 1 2, b) and T.

Linear Algebra and Dirac Notation, Pt. 3

Computation of the scattering amplitude in the spheroidal coordinates

Sample Quantum Chemistry Exam 2 Solutions

Uncorrelatedness and Independence

Estimating functional uncertainty using polynomial chaos and adjoint equations

Stochastic Processes. A stochastic process is a function of two variables:

From random matrices to free groups, through non-crossing partitions. Michael Anshelevich

Numerical Methods with MATLAB

Considering our result for the sum and product of analytic functions, this means that for (a 0, a 1,..., a N ) C N+1, the polynomial.

Lecture 45: The Eigenvalue Problem of L z and L 2 in Three Dimensions, ct d: Operator Method Date Revised: 2009/02/17 Date Given: 2009/02/11

Today: Fundamentals of Monte Carlo

Transcription:

Lecture 3: Central Limit Theorem Scribe: Jacy Bird (Division of Engineering and Applied Sciences, Harvard) February 8, 003 The goal of today s lecture is to investigate the asymptotic behavior of P N ( x) for large N. We use Laplace s Method to show that a well-behaved random variable tends to a multivariate normal distribution. This lecture also introduces normalized cumulants and briefly describes how these affect the distribution shape. General Distribution As we saw in previous lecture, the probability density function for the random-walk position after N independent steps, X N, is given by ( N ) e ik x d d k pˆ n (k) (π) d. () n= In the case where p(x) is iid, Equation () simplifies to e ik x (ˆp(k)) N d d k (π) d. () Here we have defined p(x) to be the PDF for each displacement and P N (x) to be the PDF for the position, X N, after N (IID) steps. Thus it follows that there are two moment generating functions (MGF) of interest: ˆp(k) is the MGF for x n and ˆP N (k) is the MGF for X N. Likewise, we can define a cumulant generating function (CGF) for steps x n and X N. As we saw in Lecture #, the CGF for steps x n is defined as ψ(k) = log ˆp(k) (3) = ic k k c k +... Since cumulants for IID random variables are additive, the cumulant generating function for X N is denoted ψ N = (k) log ˆP N (k) = Nψ(k). As shown in Lecture #, the variance of the distribution, σ N, equals C,N = Nc which is equivalent to Nσ. Thus the width of the distribution, σ N, scales like N. Shape of the Distribution for Large N The exact expression for P N (x) becomes increasingly complex as N increases; however, for sufficiently large N the distribution tends to a general shape. In the last section (and last lecture),

M. Z. Bazant 8.366 Random Walks and Diffusion Lecture 3 p hat (k) p hat (k) N 0.8 Increasing N 0.6 0.4 0. 0.5 0.5 0 0.5.5 k ε Figure : The dominant contribution of the integral occurs within a region ε away from the maximum point as N. it was showed that the width of this distribution scales as N. In this section we are going to investigate the N scaling more formally and use Laplace s Method to derive the asymptotic distribution shape. For a review of Asymptotic Methods, the reader is referred to Bender and Orszag (978). Laplace s Method takes advantage of the fact that the dominant contribution of an integral f(k)e Nφ(x,k) dk for large N occurs a small distance, ε, from the max φ(k) (Fig. ). We run into potential problems if we let x and N both go to infinity, as will be discussed in a later lecture. Therefore for the time being, we will only look at the central region of the distribution. Let s begin by rewriting Equation () in terms of the cumulant generating function, e ik x+nψ(k) allk d d k (π) d (4) For large N, the dominant contribution of the integral is near the origin, so we can collapse the limits of integration to a small region around the origin (this approximation introduces exponentially small error). Thus for large N, P N (x) k <ɛ e ik x e Nψ(k) d d k (π) d (5) The next step is to Taylor expand ψ(k) about the origin which as was done in Equation (3). This leads to P N (x) e ik x e inc k N k c d k+... d k (π) d (6) k <ɛ See Bender and Orszag, Advanced Mathematical Methods for Scientists and Engineers, Sec. 6.4 (Springer, 978), for a detailed discussion on Laplace s Method.

M. Z. Bazant 8.366 Random Walks and Diffusion Lecture 3 3 For sufficiently large N we can truncate the series after the k c k term. Now we expand the limits of integration back to to (again, this only introduces exponentially small errors). Thus for sufficiently large N, P N (x) e ik x e inc k N k c d k d k allk (π) d (7) e ik (x Nc ) e k (Nc ) k d d k (π) d (8) The next step is to transform this expression into a Gaussian integral. More specifically we want to transform e k (Nc ) k to e w /. Provided that c is positive definite and symmetric, we define w Nc k (9) Note : It s worth reviewing what is meant by the square root of a matrix (or nd order tensor). From linear algebra we know that M can be decomposed into QAQ T, where Q is orthogonal (Q = Q T, Q = ) and A is diagonal (A ij = a i δ ij with a i > 0 being the eigenvalues of M). In general, we can define M raised to the α power as M α QA α Q T. Here A α is diagonal with (A α ) ij = a α i δ ij. Hence, the square root of M is defined above with α =. Returning to the distribution, we make the change of variable w = Nc k (0) d d w = Nc d d k () = N d/ c d d k () In these equations c is the determinant of c and equals d i= ai where a i are the eigenvalues of c. With the substitution, Equation (8) is rewritten as P N (x) ) e i(x Nc Nc w e w (π) d N d d d w c (3) Equation (3) can be further simplified by defining a new variable, z x Nc c P N (x) c (π) d N d e iz w e w d d w (4) The above equation [Equation (4)] is expressed in terms of a multivariate Gaussian distribution. However, this distribution can also be interpreted as a product of d -D Gaussian distributions taken separately. Thus Equation (4) can be rewritten as P N (x) N d ( c e iz e w dw π ) ( e iz e w ) dw... (5) π

M. Z. Bazant 8.366 Random Walks and Diffusion Lecture 3 4 Noting that integrals in the above equation are inverse Fourier Transforms of the Gaussian e w, we simplify the expression to z P N (x) c e z e... (6) π π N d P N (x) e z (πn) d c (7) The transform to z can also be viewed as the definition of a new random variable. In the last few steps, we have actually shown that the scaled random variable Z N = (X Nc )(Nc ) = (XN X N ) c /,N (8) has a PDF, φ N (z), that tends to a multivariate standard normal distribution φ N (z) e z /. (9) (π) d/ This result is known as the Multidimensional Central Limit Theorem (CLT). Note : P N (x) d d x = φ N (z) d d z Note 3: For d =, Z N = X Nc σ since c N = σ. The CLT only applies to the central region of P N (x) where z = O(). This implies that in multiple dimensions vcx N = Nc + O( Nc ) which reduces to X N = Nc + O(σ N) when d =. Note 4: Note: An explicit assumption throughout has been the existence of c and c. If c is not finite, the CLT does not hold. Note 5: For a more rigorous derivation of the CLT, the reader is referred to W. Feller, An Introduction to Theory of Probability. Vol. I,II (970). Gnedenko, Kolmogorov (968). Note 6: Equation (9) is an asymptotic solution and it appears that the mean, c, and variance, c, of the distribution are of importance. However as we will see in the next section and in subsequent lectures, higher cumulants become important when we either look at smaller N or if we investigate the tails of the distribution.

M. Z. Bazant 8.366 Random Walks and Diffusion Lecture 3 5 0.6 0.5 Skewness Standard Normal λ 3 > 0 λ 3 < 0 0.6 0.5 Kurtosis Standard Normal λ 4 > 0 λ 4 < 0 0.4 0.4 0.3 0.3 φ(z) φ(z) 0. 0. 0. 0. 0 0 0. 4 3 0 3 4 Z 0. 4 3 0 3 4 Z Figure : Effect of skewness on distribution shape. Figure 3: shape. Effect of kurtosis on distribution Convergence in the Central Region For simplicity, let s consider d=, though it is not difficult to generalize to higher dimensions. We begin again with the distribution e ikx Nψ(k) dk e π We apply Laplace s Method to higher orders of asymptotic expansion as N (Z N remains fixed). The first step is to note that the dominate contribution to the integral occurs near k = 0, so we taylor expand ψ(k) about the origin producing ɛ ɛ e ikx e N[ ic k c k + i 3! c 3k 3 + 4! c 4k 4 +...] dk π where the imaginary numbers result from the definition of the cumulants. Substituting w = σ Nk and z = x Nc σ, the above equation reduces to N φ N (z) ɛ ɛ e iωz e ω λ 3 (iω) 3 e 3! + λ 4 N (iω) 4 4! N (0) + dω π, () where λ 3 = c 3, λ σ 3 4 = c 4,... λ σ 4 m = cm σ are the normalized cumulants of p(x). The normalized m cumulants play a large role in determining the distribution shape (Fig. and Fig. 3). For sufficiently large N, we can truncate the Taylor series and then expand the limits of integration back to to. That is, as N [ φ N (z) e iωz e ω + λ 3 (iω) 3 + λ 4 (iω) 4 3! N 4! N + ( ) λ3 (iω) 6 ( ) ] ω 5 3! N + O N dω N π. () The topics in the last part of this lecture where presented more thoroughly in 003. readers benefit, I have included part of the 003 notes by David Vener. To the

M. Z. Bazant 8.366 Random Walks and Diffusion Lecture 3 6 Noticing that allows us to rewrite Equation () as φ N (z) e z [ + λ 3 π 3! (iω) n e iωz e ω = ( ) n d n dz n e iωz e ω { H 3 (z) + λ 4 N N 4! H 4 (z) + where H n (z) is the n th Hermite polynomial defined by d n dz n e z ( ) n H n (z) e z. ( ) ( ) λ3 H 6 (z)} ] z 5 + O 3! N, (3) N Note 7: The Hermite polynomials come up in many different places in mathematical physics, including, perhaps most prominently, in the eigenfunctions for the quantum harmonic oscillator. The first few Hermite polynomials are H = z, H = z, H 3 = z 3 3z, H 4 = z 4 6z + 3, H 5 = z 5 0z 3 + 5z, H 6 = z 6 5z 4 + 5z 5. where Therefore, as N we can write Equation (3) as φ N (z) e z [ + h 3 (z) + h ] 4 (z) π N N +, (4) h 3 (z) = λ 3 3! H 3 (z), (5) h 4 (z) = λ 4 4! H 4 (z) + ( ) λ3 H 6 (z). (6) 3! Note 8: In dimension, this general expansion related to the CLT is often called the Edgeworth expansion in the mathematical literature. (See, e.g. W. Feller, An Introduction to Probability, Vol..) In more dimensions when related to random flights, expansions of this type are sometimes called Gram-Charlier Expansions. (See, e.g. B. Hughes, Random Walks and Random Environments.)