Poisson approximations

Similar documents
1 Generating functions

Tom Salisbury

Binomial and Poisson Probability Distributions

Conditional distributions (discrete case)

18.175: Lecture 17 Poisson random variables

Common Discrete Distributions

Lecture 12. Poisson random variables

POISSON PROCESSES 1. THE LAW OF SMALL NUMBERS

Special distributions

Things to remember when learning probability distributions:

(It's not always good, but we can always make it.) (4) Convert the normal distribution N to the standard normal distribution Z. Specically.

Introduction to Probability 2017/18 Supplementary Problems

Stat 134 Fall 2011: Notes on generating functions

Topic 3: The Expectation of a Random Variable

Lecture 16. Lectures 1-15 Review

Mathematical Statistics 1 Math A 6330

18.175: Lecture 13 Infinite divisibility and Lévy processes

Lecture 3. Discrete Random Variables

n px p x (1 p) n x. p x n(n 1)... (n x + 1) x!

SDS 321: Introduction to Probability and Statistics

Discrete Random Variables

1 Presessional Probability

X = X X n, + X 2

Basic Probability. Introduction

Suppose that you have three coins. Coin A is fair, coin B shows heads with probability 0.6 and coin C shows heads with probability 0.8.

A Probability Primer. A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes.

THE QUEEN S UNIVERSITY OF BELFAST

1 Random Variable: Topics

Poisson processes Overview. Chapter 10

STAT2201. Analysis of Engineering & Scientific Data. Unit 3

Topic 9 Examples of Mass Functions and Densities

SOR201 Solutions to Examples 3

Introduction to Probability

ACM 116: Lectures 3 4

On bounds in multivariate Poisson approximation

Lectures on Elementary Probability. William G. Faris

Discrete Distributions

Generating and characteristic functions. Generating and Characteristic Functions. Probability generating function. Probability generating function

Math/Stat 352 Lecture 8

The central limit theorem

2. Suppose (X, Y ) is a pair of random variables uniformly distributed over the triangle with vertices (0, 0), (2, 0), (2, 1).

6.1 Moment Generating and Characteristic Functions

18.440: Lecture 19 Normal random variables

MAS113 Introduction to Probability and Statistics. Proofs of theorems

CME 106: Review Probability theory

Lecture 10: Probability distributions TUESDAY, FEBRUARY 19, 2019

Random Models. Tusheng Zhang. February 14, 2013

Physics 403 Probability Distributions II: More Properties of PDFs and PMFs

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

3. DISCRETE RANDOM VARIABLES

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems

Discrete Mathematics for CS Spring 2007 Luca Trevisan Lecture 20

Probability Review. Gonzalo Mateos

Expectation of Random Variables

Lecture 2 Sep 5, 2017

Math 564 Homework 1. Solutions.

The expansion of random regular graphs

APPM/MATH 4/5520 Solutions to Exam I Review Problems. f X 1,X 2. 2e x 1 x 2. = x 2

Discrete Mathematics and Probability Theory Fall 2014 Anant Sahai Note 15. Random Variables: Distributions, Independence, and Expectations

Lecture 2: Discrete Probability Distributions

Stationary independent increments. 1. Random changes of the form X t+h X t fixed h > 0 are called increments of the process.

MATH 3510: PROBABILITY AND STATS June 15, 2011 MIDTERM EXAM

Lecture Notes 2 Random Variables. Discrete Random Variables: Probability mass function (pmf)

Topic 3: The Expectation of a Random Variable

Math/Stat 352 Lecture 10. Section 4.11 The Central Limit Theorem

SUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416)

Probability: Why do we care? Lecture 2: Probability and Distributions. Classical Definition. What is Probability?

More on Distribution Function

CSE 103 Homework 8: Solutions November 30, var(x) = np(1 p) = P r( X ) 0.95 P r( X ) 0.

Lecture 2: Probability and Distributions

Chapter 3 Discrete Random Variables

MAS113 Introduction to Probability and Statistics. Proofs of theorems

S n = x + X 1 + X X n.

Lecture 1: Review on Probability and Statistics

Probability Review. Yutian Li. January 18, Stanford University. Yutian Li (Stanford University) Probability Review January 18, / 27

Relationship between probability set function and random variable - 2 -

Probabilistic Systems Analysis Spring 2018 Lecture 6. Random Variables: Probability Mass Function and Expectation

On the Entropy of Sums of Bernoulli Random Variables via the Chen-Stein Method

Introduction to Probability

Chapter 8. Some Approximations to Probability Distributions: Limit Theorems

3. Review of Probability and Statistics

Guidelines for Solving Probability Problems

Notes on Poisson Approximation

Single Maths B: Introduction to Probability

Advanced Herd Management Probabilities and distributions

Why study probability? Set theory. ECE 6010 Lecture 1 Introduction; Review of Random Variables

Lectures 16-17: Poisson Approximation. Using Lemma (2.4.3) with θ = 1 and then Lemma (2.4.4), which is valid when max m p n,m 1/2, we have

CSE 312 Final Review: Section AA

Lecture 2: Review of Basic Probability Theory

1 INFO Sep 05

4th IIA-Penn State Astrostatistics School July, 2013 Vainu Bappu Observatory, Kavalur

Lecture 13. Poisson Distribution. Text: A Course in Probability by Weiss 5.5. STAT 225 Introduction to Probability Models February 16, 2014

MFM Practitioner Module: Quantitative Risk Management. John Dodson. September 23, 2015

Chapter 2: Discrete Distributions. 2.1 Random Variables of the Discrete Type

Discrete Distributions Chapter 6

1.1 Review of Probability Theory

Example continued. Math 425 Intro to Probability Lecture 37. Example continued. Example

Review of Probability Theory

Bipartite decomposition of random graphs

Bernoulli and Binomial

Transcription:

Chapter 9 Poisson approximations 9.1 Overview The Binn, p) can be thought of as the distribution of a sum of independent indicator random variables X 1 + + X n, with {X i = 1} denoting a head on the ith toss of a coin that lands heads with probability p. Each X i has a Berp) distribution. The normal approximation to the Binomial works best when the variance np1 p) is large, for then each of the standardized summands X i p)/ np1 p) makes a relatively small contribution to the standardized sum. When n is large but p is small, in such a way that λ := np is not too large, a different type of approximation to the Binomial is better. The traditional explanation uses an approximation to ) n P{X = k} = p k 1 p) n k k for a fixed k. For a k that is small compared with n, consider the contributions n k) p k and 1 p) n k separately. and n k ) p k = nn 1)... n k + 1) = 1 1 1 )... n ) λ k n 1 k 1 n ) λ k λk log1 p) n k = n k) log 1 λ/n) n λ/n). 1

9. Poisson approximations 2 That is, 1 p) n k e λ. Together the two approximations give n )p k 1 p) n k λk k e λ. For large k, both P{X = k} and p k := e λ λ k / are small. The p k define a new distribution. Definition. A random variable Y is said to have a Poisson distribution with parameter λ if it can take values in N 0, the set of nonnegative integers, with probabilities P{Y = k} = e λ λ k for k = 0, 1, 2,... The parameter λ must be positive. The distribution is denoted by Poissonλ). That is, for λ = np not too large, the Binn, p) is well?) approximated by the Poissonλ). Remark. Counts of rare events such as the number of atoms undergoing radioactive decay during a short period of time, or the number of aphids on a leaf are often modeled by Poisson distributions, at least as a first approximation. The Poisson inherits several properties from the Binomial. For example, the Binn, p) has expected value np and variance np1 p). One might suspect that the Poissonλ) should therefore have expected value λ = nλ/n) and variance λ = lim n nλ/n)1 λ/n). Also, the coin-tossing origins of the Binomial show that if X has a Binm, p) distribution and Y has a Binn, p) distribution independent of X, then X + Y has a Binn + m, p) distribution. Putting λ = mp and µ = np one might then suspect that the sum of independent Poissonλ) and Poissonµ) distributed random variables is Poissonλ + µ) distributed. These suspicions are correct. Example <9.1> If X has a Poissonλ) distribution, then EX = varx) = λ. If also Y has a Poissonµ) distribution, and Y is independent of X, then X + Y has a Poissonλ + µ) distribution. There is a clever way to simplify some of the calculations in the last Example using generating functions, a way to code all the Poisson probabilities into a single function on [0, 1].

9. Poisson approximations 3 Example <9.2> Calculate moments of the Poissonλ) distribution using its generating function. 9.2 A more precise Poisson approximation Modern probability methods have improved this rough approximation of the Binomial by the Poisson by giving useful upper bounds for the error of approximation. Using a technique known as the Chen-Stein method one can show that d T V Binn, p), Poissonnp)) := 1 2 k 0 P{S = k} e λ λk min p, np 2), which makes precise the traditional advice that the Poisson approximation is good when p is small and np is not too big. In fact, the tradition was a bit conservative.) Remark. The quantity d T V P, Q) is called the total variation distance between two probabilities P and Q. It is also equal to max A P A QA where the maximum runs over all subsets A of the set where both P and Q are defined. For P = Binn, p) and Q = Poissonnp), the A runs over all subsets of the nonnegative integers. The Chen-Stein method of approximation also works in situations where the rare events do not all have the same probability of occurrence. For example, suppose S = X 1 + X 2 + + X n, a sum of independent random variables where X i has a Berp i ) distribution, for constants p 1, p 2,..., p n that are not necessarily all the same. The sum S has expected value λ = p 1 + + p n. Using Chen-Stein it can also be shown that that 1 2 k 0 P{S = k} e min 1, 1 ) n λ i=1 p2 i. λ λk The Chen-Stein method of proof is elementary in the sense that it makes use of probabilistic techniques only at the level of Statistics 241 but extremely subtle. See Barbour et al. 1992) for an extensive discussion of the method.

9. Poisson approximations 4 9.3 Poisson approximations under dependence The Poisson approximation also applies in many settings where the trials are almost independent, but not quite. Again the Chen-Stein method delivers impressively good bounds on the errors of approximation. For example, the method works well in two cases where the dependence takes an a simple form. Once again suppose S = X 1 + X 2 + + X n, where X i has a Berp i ) distribution, for constants p 1, p 2,..., p n that are not necessarily all the same. Often X i is interpreted as the indicator function for success in the ith in some finite set of trials. Define S i = S X i = 1 j n I{j i}x j. The random variables X 1,..., X n are said to be positively associated if P{S i k X i = 1} P{S i k X i = 0} for each i and k = 0, 1, 2,... and negatively associated if P{S i k X i = 1} P{S i k X i = 0} for each i and k = 0, 1, 2,.... Intuitively, positive association means that success in the ith trial makes success in the other trials more likely; negative association means that success in the ith trial makes success in the other trials less likely. With some work it can be shown Barbour et al., 1992, page 20) that 1 2 1, 1 ) λ λ λk P{S = k} e min k 0 { vars) λ + 2 n ) i=1 p2 i under positive association λ vars)) under negative association. These bounds take advantage of the fact that vars) would be exactly equal to λ if S had a Poissonλ) distribution. The next Example illustrates both the classical approach and the Chen- Stein approach via positive association) to deriving a Poisson approximation for a matching problem. Example <9.3> Poisson approximation for a matching problem: assignment of n letters at random to n envelopes, one per envelope.

9. Poisson approximations 5 9.4 Examples for Chapter 9 <9.1> Example. If X has a Poissonλ) distribution, then EX = varx) = λ. If also Y has a Poissonµ) distribution, and Y is independent of X, then X +Y has a Poissonλ + µ) distribution. Assertion i) comes from a routine application of the formula for the expectation of a random variable with a discrete distribution. EX = kp{x = k} = k e λ λ k What happens to k = 0? k=0 k=1 = e λ λ k 1 λ k 1)! = e λ λe λ = λ. k 1=0 Notice how the k cancelled out one factor from the in the denominator. If I were to calculate EX 2 ) in the same way, one factor in the k 2 would cancel the leading k from the, but would leave an unpleasant k/k 1)! in the sum. Too bad the k 2 cannot be replaced by kk 1). Well, why not? EX 2 X) = kk 1)P{X = k} k=0 = e λ k=2 = e λ λ 2 Now calculate the variance. kk 1) λk k 2=0 λ k 2 k 2)! = λ2. What happens to k = 0 and k = 1? varx) = EX 2 ) EX) 2 = EX 2 X) + EX EX) 2 = λ. For assertion iii), first note that X + Y can take only values 0, 1, 2.... For a fixed k in this range, decompose the event {X + Y = k} into disjoint pieces whose probabilities can be simplified by means of the independence

9. Poisson approximations 6 between X and Y. P{X + Y = k} = P{X = 0, Y = k} + P{X = 1, Y = k 1} + + P{X = k, Y = 0} = P{X = 0}P{Y = k} + + P{X = k}p{y = 0} = e λ λ 0 0! = e λ µ = e λ µ e µ µ k + + e λ λ k e µ µ 0 0! 0! λ0 µ k + 1!k 1)! λ1 µ k 1 + + 0! λk µ 0 λ + µ) k. The bracketed sum in the second last line is just the binomial expansion of λ + µ) k. Remark. How do you interpret the notation in the last calculation when k = 0? I always feel slightly awkward about a contribution from k 1 if k = 0. <9.2> Example. There is a sneakier way to calculate EX m for m = 1, 2,... when X has a Poissonλ) distribution. Code the whole distribution into a function the probability generating function) of a dummy variable s: gs) := Es X = k 0 sk λ λk e = sλ) k e λ = e λ e λs. k 0 Given g, the individual probabilities P{X = k} could be recovered by expanding the function as a power series in s. Other facts about the distribution can also be obtained from g. For example, d gs) = lim ds E h 0 s + h) X s X ) = E h s sx = EXs X 1 and, by direct calculation, g s) = e λ λe λs. Put s = 1 in both expressions to deduce that EX = g 1) = λ. Similarly, repeated differentiation inside the expectation sign gives g m) s) = m s m EsX ) = E XX 1)... X m + 1)s X m), and direct differentiation of g gives g m) s) = e λ λ m e λs. Again put s = 1 to deduce that λ m = g m) 1) = E XX 1)... X m + 1) ) for m = 1, 2,... )

9. Poisson approximations 7 <9.3> Example. Suppose n letters are placed at random into n envelopes, one letter per envelope. The total number of correct matches, S, can be written as a sum X 1 + + X n of indicators, { 1 if letter i is placed in envelope i, X i = 0 otherwise. The X i are dependent on each other. For example, symmetry implies that p i = P{X i = 1} = 1/n for each i and P{X i = 1 X 1 = X 2 = = X i 1 = 1} = 1 n i + 1 Remark. If we eliminated the dependence by relaxing the requirement of only one letter per envelope, the number of letters placed in the correct envelope possibly together with other, incorrect letters) would then have a Binn, 1/n) distribution, which is approximated by Poisson1) if n is large. We can get some supporting evidence for S having something close to a Poisson1) distribution under the original assumption one letter per envelope) by calculating some moments. ES = i n EX i = np{x i = 1} = 1 and ES 2 = E X 1 2 + + Xn 2 + 2 X i X j i<j ) n = nex1 2 + 2 EX 1 X 2 2 by symmetry = np{x 1 = 1} + n 2 n)p{x 1 = 1, X 2 = 1} = n 1 ) + n 2 1 n) n nn 1) = 2. Thus vars) = ES 2 ES) 2 = 1. Compare with Example <9.1>, which gives EY = 1 and vary ) = 1 for a Y distributed Poisson1).

9. Poisson approximations 8 Using the method of inclusion and exclusion, it is possible Feller, 1968, Chapter 4) to calculate the exact distribution of the number of correct matches, P{S = k} = 1 1 1 1! + 1 2! 1 ) 3! ± 1 for k = 0, 1,..., n. n k)! For fixed k, as n the probability converges to 1 1 1 + 12! 13! )... = e 1, which is the probability that Y = k if Y has a Poisson1) distribution. The Chen-Stein method is also effective in this problem. I claim that it is intuitively clear although a rigorous proof might be tricky) that the X i s are positively associated: P{S i k X i = 1} P{S i k X i = 0} for each i and each k N 0. I feel that if X i = 1, then it is more likely for the other letters to find their matching envelopes than if X i = 0, which makes things harder by filling one of the envelopes with the incorrect letter i. Positive association gives 1 2 k 0 P{S = k} e λ λk 2 n i=1 p2 i + vars) 1 = 2/n. As n gets large, the distribution of S does get close to the Poisson1) in the strong, total variation sense. References Barbour, A. D., L. Holst, and S. Janson 1992). Poisson Approximation. Oxford University Press. Feller, W. 1968). An Introduction to Probability Theory and Its Applications third ed.), Volume 1. New York: Wiley.