Lecture 7: Statistics and the Central Limit Theorem. Philip Moriarty,

Similar documents
Lecture 8: Differential Equations. Philip Moriarty,

Getting Started with Communications Engineering

ACCESS TO SCIENCE, ENGINEERING AND AGRICULTURE: MATHEMATICS 1 MATH00030 SEMESTER / Lines and Their Equations

MA 1125 Lecture 15 - The Standard Normal Distribution. Friday, October 6, Objectives: Introduce the standard normal distribution and table.

How much does the cow weigh?

More on infinite series Antiderivatives and area

Science One Physics Work and Energy

DIFFERENTIAL EQUATIONS

Computational Techniques Prof. Dr. Niket Kaisare Department of Chemical Engineering Indian Institute of Technology, Madras

Tips and Tricks in Real Analysis

Math Lab 10: Differential Equations and Direction Fields Complete before class Wed. Feb. 28; Due noon Thu. Mar. 1 in class

Lecture Notes Set 3a: Probabilities, Microstates and Entropy

Knots, Coloring and Applications

Making Measurements. On a piece of scrap paper, write down an appropriate reading for the length of the blue rectangle shown below: (then continue )

Random number generation

Confidence Intervals

Conservation of Momentum

Math Lecture 23 Notes

Resonance and response

STA Why Sampling? Module 6 The Sampling Distributions. Module Objectives

Joint Probability Distributions and Random Samples (Devore Chapter Five)

First Derivative Test

Around the corner. Mathematics B-day 2015, Friday November 13, 9:00h-16:00h

Constrained and Unconstrained Optimization Prof. Adrijit Goswami Department of Mathematics Indian Institute of Technology, Kharagpur

MITOCW watch?v=rf5sefhttwo

Vectors Part 1: Two Dimensions

carroll/notes/ has a lot of good notes on GR and links to other pages. General Relativity Philosophy of general

Constructing Potential Energy Diagrams

Introduction to Machine Learning Prof. Sudeshna Sarkar Department of Computer Science and Engineering Indian Institute of Technology, Kharagpur

Expected Value II. 1 The Expected Number of Events that Happen

ECE Homework Set 2

22 Approximations - the method of least squares (1)

MA 1128: Lecture 08 03/02/2018. Linear Equations from Graphs And Linear Inequalities

Thermodynamics (Classical) for Biological Systems. Prof. G. K. Suraishkumar. Department of Biotechnology. Indian Institute of Technology Madras

PHYSICS 15a, Fall 2006 SPEED OF SOUND LAB Due: Tuesday, November 14

COMP6053 lecture: Sampling and the central limit theorem. Markus Brede,

Q: How can quantum computers break ecryption?

Chapter 1: Useful definitions

Dot Products, Transposes, and Orthogonal Projections

Probability Methods in Civil Engineering Prof. Rajib Maity Department of Civil Engineering Indian Institute of Technology, Kharagpur

Linear Algebra, Summer 2011, pt. 2

Solving with Absolute Value

( )( b + c) = ab + ac, but it can also be ( )( a) = ba + ca. Let s use the distributive property on a couple of

Finding Limits Graphically and Numerically

Electromagnetic Theory Prof. D. K. Ghosh Department of Physics Indian Institute of Technology, Bombay

Notes on Random Variables, Expectations, Probability Densities, and Martingales

ARE202A, Fall 2005 CONTENTS. 1. Graphical Overview of Optimization Theory (cont) Separating Hyperplanes 1

6. Qualitative Solutions of the TISE

Digital Signal Processing Prof. T. K. Basu Department of Electrical Engineering Indian Institute of Technology, Kharagpur

Problem Solving. Kurt Bryan. Here s an amusing little problem I came across one day last summer.

Chapter 8: Trig Equations and Inverse Trig Functions

A Series Transformations

1 Normal Distribution.

Overview. Probabilistic Interpretation of Linear Regression Maximum Likelihood Estimation Bayesian Estimation MAP Estimation

Change & Rates of Change

!P x. !E x. Bad Things Happen to Good Signals Spring 2011 Lecture #6. Signal-to-Noise Ratio (SNR) Definition of Mean, Power, Energy ( ) 2.

Chapter 18. Sampling Distribution Models /51

Lesson 3-1: Solving Linear Systems by Graphing

Guide to Proofs on Sets

Propositional Logic: Semantics and an Example

Engineering Mechanics Prof. U. S. Dixit Department of Mechanical Engineering Indian Institute of Technology, Guwahati Kinematics

Experiment 2: Projectile Motion

Random Variate Generation

has a lot of good notes on GR and links to other pages. General Relativity Philosophy of general relativity.

Basic Quantum Mechanics Prof. Ajoy Ghatak Department of Physics Indian Institute of Technology, Delhi

Instructor (Brad Osgood)

Lecture 6. Statistical Processes. Irreversibility. Counting and Probability. Microstates and Macrostates. The Meaning of Equilibrium Ω(m) 9 spins

Designing geared 3D twisty puzzles (on the example of Gear Pyraminx) Timur Evbatyrov, Jan 2012

Flat Geometry. Spherical Geometry

PHYS 7411 Spring 2015 Computational Physics Homework 4

Physics 342 Lecture 2. Linear Algebra I. Lecture 2. Physics 342 Quantum Mechanics I

One-sample categorical data: approximate inference

Physics E-1ax, Fall 2014 Experiment 3. Experiment 3: Force. 2. Find your center of mass by balancing yourself on two force plates.

UNIT Define joint distribution and joint probability density function for the two random variables X and Y.

Statistical Distribution Assumptions of General Linear Models

Lecture 8 : Structural Induction DRAFT

Experiment 1: The Same or Not The Same?

/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Matroids and Greedy Algorithms Date: 10/31/16

Forces as Interactions

Modern Algebra Prof. Manindra Agrawal Department of Computer Science and Engineering Indian Institute of Technology, Kanpur

Lecture 9: Predictive Inference for the Simple Linear Model

Lecture 06 Lecture 4: Part II - Uncertainty principle

POLYNOMIAL EXPRESSIONS PART 1

Quantum Mechanics - I Prof. Dr. S. Lakshmi Bala Department of Physics Indian Institute of Technology, Madras. Lecture - 9 Introducing Quantum Optics

CALCULUS III. Paul Dawkins

When a function is defined by a fraction, the denominator of that fraction cannot be equal to zero

COMP6053 lecture: Sampling and the central limit theorem. Jason Noble,

Math101, Sections 2 and 3, Spring 2008 Review Sheet for Exam #2:

Predator - Prey Model Trajectories and the nonlinear conservation law

Module 2: Reflecting on One s Problems

IE 361 Module 4. Modeling Measurement. Reading: Section 2.1 Statistical Methods for Quality Assurance. ISU and Analytics Iowa LLC

Basic Quantum Mechanics Prof. Ajoy Ghatak Department of Physics Indian Institute of Technology, Delhi

Functions. If x 2 D, then g(x) 2 T is the object that g assigns to x. Writing the symbols. g : D! T

TEMPORAL EXPONENTIAL- FAMILY RANDOM GRAPH MODELING (TERGMS) WITH STATNET

Error Correcting Codes Prof. Dr. P. Vijay Kumar Department of Electrical Communication Engineering Indian Institute of Science, Bangalore

ASTRO 114 Lecture Okay. What we re going to discuss today are what we call radiation laws. We ve

SPH4U1 Lesson 02 Measurement & Graphing

Series, Parallel, and other Resistance

Operation and Supply Chain Management Prof. G. Srinivasan Department of Management Studies Indian Institute of Technology, Madras

Transcription:

Lecture 7: Statistics and the Central Limit Theorem Philip Moriarty, philip.moriarty@nottingham.ac.uk NB Notes based heavily on lecture slides prepared by DE Rourke for the F32SMS module, 2006

7.1 Recap and Overview In Lectures 1 6 we have seen how Fourier analysis is at the core of a variety of fields in physics (and science in general) with a particular recent focus on signal and image analysis/processing. In the final two lectures of the F32SMS module we will examine how Fourier transforms (FTs) can be used (perhaps surprisingly) in two areas rather different to those studied in this module to date: statistics and the solution of differential equations. Today we will see how Fourier analysis can solve a central problem in statistics the derivation of the central limit theorem. In doing so you will see why the Gaussian distribution is so prevalent in Nature and find a remarkable link between convolution and the statistics of random processes. 7.2 Probability Density Functions You will be familiar with the concept of a probability density function from the 1 st year Thermal & Kinetic module and your 1 st and 2 nd year quantum mechanics-related modules. Let s start with considering the distributions produced by perhaps the simplest random number generator: dice.? Sketch the probability density function you would expect for the numbers generated by the throw of a single die.? Sketch the probability density function you would expect for the numbers generated by the throw of two dice.

Let s try another example. We buy a random apple from a supermarket and we weigh it. The weight is an example of a random variable. The apple is unlikely to weigh exactly 1 kg or 1g there will be a most probable mass. The probabilities for the mass are reflected in the probability density function (PDF), φ(x), shown in Fig. 1 below. Fig. 1 Probability density function for the mass of an apple. The area under the φ(x) curve between, for example, a and b gives the probability, p, that the mass takes a value between a and b. That is, This of course means that the total area under the φ(x) curve, if the function is appropriately normalised, must be 1, since the probability of x lying between - and + must be 1.? Given a normalised probability density function φ(x), how would you find the mean value of x, i.e. <x>? Similarly, how would you find the variance, σ 2, where σ 2 = <(x-<x>)> 2?

The property that is taken as the defining property of a PDF is : Now, for simplicity, let s assume that the PDF for the weights of the apples is the uniform distribution (i.e. like that for the single die example) shown in Fig. 2 below. Fig. 2 A uniform probability distribution function. This means that no apples weigh less than 90g or more than 110 g and that any weight between 90 and 110 is equally likely. The average mass is clearly 100 g and, perhaps less clearly, the standard deviation, σ, is a/ 3 where a is the half-width of the distribution.? Why is the height of the function 0.05 g -1? Again, see the 1 st year Thermal & Kinetic notes (available at www.nottingham.ac.uk/~ppzpjm/f31st1)

Now, suppose we buy two apples and calculate the sum of the two masses, S = x 1 + x 2. What is the PDF for the sum, S? As each individual apple weighs between 90g and 110 g, there s no way the sum can be less than 180g or more than 220g, so the PDF must be zero for S<180g or S>220g. Similarly, what is the PDF for the sum of the masses if we buy three apples, four apples, or, in general, n apples? What happens in the limit n? These are the questions we ll address in this lecture. A very important result called the central limit theorem tells us what happens in the n limit. Importantly, while we ll deal with apples, you ll not be surprised to know that the central limit theorem is not restricted to the analysis of fruit! It is a general theorem which is applicable, in particular, to the measurement of quantities in an experiment. If we try to measure a quantity, Q, in an experiment, the measured value will on the whole equal Q plus random noise. If the measurement is repeated n times (assuming n is large) and the values summed together, one would expect to average out the noise, and so the sum calculated will be near to nq. This means that the PDF should be strongly peaked at nq.

7.3 Numerical experiments PDFs with Matlab Fig.3 below shows the results of a Matlab program to generate random numbers between 90 and 110. Note that the random number generator produces a flat, i.e. uniform distribution. Fig. 3 A uniform probability distribution function generated by Matlab? From what we ve done before, perhaps you can guess what the PDF for the sum, S, of two random numbers (x 1 and x 2 ) should look like? and as we increase the number of random values we use in the sum, the distribution function approaches a limiting (and hopefully very familiar) form:

7.4 Analytically calculating PDFs Although numerical experiments provide a lot of insight, it is of course much more elegant (and efficient) if we can analytical derive an equation for the PDF. Let s look at the two apple case first of all, as shown in Fig. 4. Fig. 4 Probability distribution function for, S, the sum of the masses of two apples. We need to bear in mind the definition of φ s (S): That is, φ s (S) ds is the probability of the total weight of two apples falling between S and S + ds. We now need to work out what this probability is. (It s best to refer to the Powerpoint slides for Lecture 7 at this point if you re not already doing so.) Imagine whenever we pick an apple of weight x 1 followed by an apple of weight x 2, we draw a dot on a 2D graph at co-ordinates x 1, x 2 (see Fig. 5(a) below). The probability of a dot landing in a square defined by the ranges x 1 x 1 + dx 1 and x 2 x 2 + dx 2 is then given by p (see Fig. 5). Fig. 5(a) Probability of choosing apple weights within given ranges.

Suppose we re interested not in x 1 and x 2 individually but in the sum, S, of x 1 and x 2. If S has some fixed value then x 1 and x 2 must lie on the line shown in Fig. 5(b) below. Fig. 5(b)

On the other hand, if the sum equals S + ds then x 1 and x 2 must lie on the line shown in Fig. 5(c) below. Fig. 5(c)

The probability of our dot landing between the two lines is the probability written above (i.e. φ(x 1 )φ(x 2 )dx 1 dx 2 ) but integrated over the total area between the two lines. For a fixed x 1, x 2 can range from S-x 1 to S+dS-x 1. Thus, we can write down the probability as follows: Although this looks a little formidable, the integral isn t too scary! The first thing we can do is move φ(x 1 ) to outside the inner integration. Although it at first appears surprising that we can do this, remember the following formula: That means we can write the integral involving x 1 and x 2 as shown on the following page.

That is, And now perhaps you start to see the link with Fourier transforms. The convolution of two functions f and g can be written as: In words, the PDF for the sum of x 1 and x 2 is the self-convolution of the PDF for a single apple s weight, φ.

The numerical experiments (and our dice games) suggested that for two apples or two dice the PDF should be triangular. This is exactly what we d expect from the convolution of two top hat functions! (See Problems Class 3 and Fig. 6 below). Fig. 6 Convolution of the single apple probability distribution functions yields the PDF for, S, the sum of the weights of two apples. It s then easy to generalise this result to the PDF for the total weight of 3 or more apples. As shown in Fig. 7, we convolve φ with itself n times to get the PDF for the total weight of n apples. Fig. 7 Use of convolution to derive PDF for weight of n apples.

7.5 PDFs and Fourier Transforms Let s now consider using Fourier transforms to carry out the convolution (i.e. let s switch to k-space (and back again)). Fig. 8 shows the Fourier transform of the PDF which, as the PDF is a top-hat function, you should know is a sinc function. (? Why is the exp(- ik<x>) term required?) If the sinc function is raised to the tenth power (i.e. if we are considering 10 apples), then the function that results is shown in the bottom right corner of Fig.8. To get the resulting PDF function we have to move from k-space and thus need to use an inverse Fourier transform. This yields the function shown in the bottom left corner of Fig. 8. Fig. 8 Convolution in Fourier space. The PDF for the total weight of ten apples is shown in the bottom left corner.

7.6 The central limit theorem. While the analytical expression for the final PDF of the total weight of ten apples is too horrible to write down, there is an approximate formula which gets more and more accurate the more apples we choose to include, i.e. the larger n is. This formula is shown in Fig. 9 and while it too perhaps looks horrible (!), it is nothing more than the Gaussian or normal distribution. This is called the central limit theorem: φ s (S) tends to a Gaussian distribution for large n. The central limit theorem is just a statement about the result of convolving a function with itself n times. It doesn t matter what the shape of the PDF is the central limit theorem tells use that if we convolve it with itself n times (where n is a large number), we ll get a Gaussian. This is a remarkable result: even if we don t know how noise is affecting our measurements, if we make enough measurements, we know that their sum must have a Gaussian distribution! Fig. 9 Upper: Equation for, and sketch of, a Gaussian distribution; Lower: A top-hat function convolved with itself 10 times produces a Gaussian distribution.