Common Discrete Distributions

Similar documents
Suppose that you have three coins. Coin A is fair, coin B shows heads with probability 0.6 and coin C shows heads with probability 0.8.

BINOMIAL DISTRIBUTION

Math/Stat 352 Lecture 8

Topic 3: The Expectation of a Random Variable

Tom Salisbury

Binomial random variable

Random Variables Example:

Relationship between probability set function and random variable - 2 -

Bernoulli and Binomial

Things to remember when learning probability distributions:

Special Mathematics Discrete random variables

Lecture 08: Poisson and More. Lisa Yan July 13, 2018

Lecture 3. Discrete Random Variables

CMPSCI 240: Reasoning Under Uncertainty

Chapter 2: Discrete Distributions. 2.1 Random Variables of the Discrete Type

a zoo of (discrete) random variables

Discrete Distributions

Introduction to Statistical Data Analysis Lecture 3: Probability Distributions

Mathematical Statistics 1 Math A 6330

(Ch 3.4.1, 3.4.2, 4.1, 4.2, 4.3)

1 Bernoulli Distribution: Single Coin Flip

Stat 134 Fall 2011: Notes on generating functions

4. Discrete Probability Distributions. Introduction & Binomial Distribution

Poisson approximations

Discrete Distributions

Lecture 13. Poisson Distribution. Text: A Course in Probability by Weiss 5.5. STAT 225 Introduction to Probability Models February 16, 2014

Binomial and Poisson Probability Distributions

Guidelines for Solving Probability Problems

MATH Notebook 5 Fall 2018/2019

Some Special Discrete Distributions

Chapter 3 Discrete Random Variables

Random Variable. Discrete Random Variable. Continuous Random Variable. Discrete Random Variable. Discrete Probability Distribution

(Ch 3.4.1, 3.4.2, 4.1, 4.2, 4.3)

Random Variables. Definition: A random variable (r.v.) X on the probability space (Ω, F, P) is a mapping

Chapters 3.2 Discrete distributions

Continuous-Valued Probability Review

n px p x (1 p) n x. p x n(n 1)... (n x + 1) x!

Statistics for Economists. Lectures 3 & 4

Lecture 12. Poisson random variables

3.4. The Binomial Probability Distribution

Special Discrete RV s. Then X = the number of successes is a binomial RV. X ~ Bin(n,p).

Page Max. Possible Points Total 100

18.175: Lecture 13 Infinite divisibility and Lévy processes

Lecture 14. Text: A Course in Probability by Weiss 5.6. STAT 225 Introduction to Probability Models February 23, Whitney Huang Purdue University

p. 4-1 Random Variables

CSE 312: Foundations of Computing II Random Variables, Linearity of Expectation 4 Solutions

Outline PMF, CDF and PDF Mean, Variance and Percentiles Some Common Distributions. Week 5 Random Variables and Their Distributions

Bayes Theorem and Hypergeometric Distribution

Topic 3: The Expectation of a Random Variable

CSE 312: Foundations of Computing II Quiz Section #10: Review Questions for Final Exam (solutions)

Topic 3 - Discrete distributions

Conditional Probability

Polytechnic Institute of NYU MA 2212 MIDTERM Feb 12, 2009

Bernoulli Trials, Binomial and Cumulative Distributions

Discrete random variables and probability distributions

Stat 100a, Introduction to Probability.

a zoo of (discrete) random variables

HW on Ch Let X be a discrete random variable with V (X) = 8.6, then V (3X+5.6) is. V (3X + 5.6) = 3 2 V (X) = 9(8.6) = 77.4.

Discrete Probability Distributions

Expectations. Definition Let X be a discrete rv with set of possible values D and pmf p(x). The expected value or mean value of X, denoted by E(X ) or

Fundamental Tools - Probability Theory II

Random Variables. Statistics 110. Summer Copyright c 2006 by Mark E. Irwin

ST 371 (V): Families of Discrete Distributions

2. In a clinical trial of certain new treatment, we may be interested in the proportion of patients cured.

Probability and Statistics for Engineers

STA 4321/5325 Solution to Extra Homework 1 February 8, 2017

Analysis of Engineering and Scientific Data. Semester

More Discrete Distribu-ons. Keegan Korthauer Department of Sta-s-cs UW Madison

Topic 9 Examples of Mass Functions and Densities

EE126: Probability and Random Processes

STAT509: Discrete Random Variable

b. ( ) ( ) ( ) ( ) ( ) 5. Independence: Two events (A & B) are independent if one of the conditions listed below is satisfied; ( ) ( ) ( )

Random variables. DS GA 1002 Probability and Statistics for Data Science.

Introduction to Probability and Statistics Slides 3 Chapter 3

Lecture 13 (Part 2): Deviation from mean: Markov s inequality, variance and its properties, Chebyshev s inequality

Random variables. Lecture 5 - Discrete Distributions. Discrete Probability distributions. Example - Discrete probability model

Lecture 2: Discrete Probability Distributions

Chapter 8: An Introduction to Probability and Statistics

Debugging Intuition. How to calculate the probability of at least k successes in n trials?

INF FALL NATURAL LANGUAGE PROCESSING. Jan Tore Lønning

Applied Statistics I

Probability Density Functions and the Normal Distribution. Quantitative Understanding in Biology, 1.2

Probability Theory and Simulation Methods. April 6th, Lecture 19: Special distributions

Lecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes

IEOR 3106: Introduction to Operations Research: Stochastic Models. Professor Whitt. SOLUTIONS to Homework Assignment 1

(b). What is an expression for the exact value of P(X = 4)? 2. (a). Suppose that the moment generating function for X is M (t) = 2et +1 3

Fundamental Tools - Probability Theory IV

Closed book and notes. 120 minutes. Cover page, five pages of exam. No calculators.

Special distributions

ELEG 3143 Probability & Stochastic Process Ch. 2 Discrete Random Variables

GEOMETRIC -discrete A discrete random variable R counts number of times needed before an event occurs

Discrete random variables

Lecture 8 : The Geometric Distribution

Lecture 2: Repetition of probability theory and statistics

STAT 414: Introduction to Probability Theory

PAS04 - Important discrete and continuous distributions

Problem 1. Problem 2. Problem 3. Problem 4

Chapter 3. Discrete Random Variables and Their Probability Distributions

Lecture 1: Review on Probability and Statistics

18.175: Lecture 17 Poisson random variables

Transcription:

Common Discrete Distributions Statistics 104 Autumn 2004 Taken from Statistics 110 Lecture Notes Copyright c 2004 by Mark E. Irwin

Common Discrete Distributions There are a wide range of popular discrete distributions used in probability modelling and statistics. A subset which will be discussed here are Bernoulli and Binomial Geometric and Negative Binomial Poisson Hypergeometric Discrete Uniform Common Discrete Distributions 1

Bernoulli and Binomial Bernoulli: This distribution is useful for describing the results of a single trial that is either a success (Prob = p) or a failure (Prob = 1 p = q). p(x) = p x = 1 1 p x = 0 0 Otherwise This is the model I ve been using for flipping a biased coin. Its also the distribution of an indicator RV. (Denoted by Ber(p)) E[X] = p; V ar(x) = p(1 p); SD(X) = p(1 p) Discrete Distributions 2

Binomial: Let Y 1, Y 2,..., Y n be independent Ber(p) RVs. Then X = n i=1 Y i is a binomial RV (Denoted Bin(n, p)). Note that Ber(p) = Bin(1, p) X is the number of successes in n independent, identical (same p) trials. It is used in a wide range of problems: Medical: number of patients, out of 500 in a trial, surviving 5 years Quality control: sample 10 parts and record the number that are faulty Coin flipping: number of heads in 50 flips. Note in this second case, what we are calling a success (what we are counting) is actually a failure for a practical purposes. Discrete Distributions 3

The PMF for the binomial distribution is p(k) = ( ) n p k (1 p) n k ; k k = 0, 1,..., n Table A in the back of the book gives the CDF for the binomial distribution for some values of n (5, 10, 15, 20,2 5) and p (0.01, 0.05, 0.10,..., 0.9, 0.95, 1.0) The following is copy of the table for n = 5 p k 0.01 0.05 0.10 0.20 0.30 0.40 0.50 0.60 0.70 0.80 0.90 0.95 0.99 0 0.951 0.774 0.590 0.328 0.168 0.078 0.031 0.010 0.002 0.000 0.000 0.000 0.000 1 0.999 0.977 0.919 0.737 0.528 0.337 0.188 0.087 0.031 0.007 0.000 0.000 0.000 2 1.000 0.999 0.991 0.942 0.837 0.683 0.500 0.317 0.163 0.058 0.009 0.001 0.000 3 1.000 1.000 1.000 0.993 0.969 0.913 0.812 0.663 0.472 0.263 0.081 0.023 0.001 4 1.000 1.000 1.000 1.000 0.998 0.990 0.969 0.922 0.832 0.672 0.410 0.226 0.049 Note that the row for k = n must be all 1, so it is omitted. Discrete Distributions 4

k p 0.01 0.05 0.10 0.20 0.30 0.40 0.50 0.60 0.70 0 0.951 0.774 0.590 0.328 0.168 0.078 0.031 0.010 0.002 1 0.999 0.977 0.919 0.737 0.528 0.337 0.188 0.087 0.031 2 1.000 0.999 0.991 0.942 0.837 0.683 0.500 0.317 0.163 3 1.000 1.000 1.000 0.993 0.969 0.913 0.812 0.663 0.472 4 1.000 1.000 1.000 1.000 0.998 0.990 0.969 0.922 0.832 The table can be used to get the following probabilities (for n = 5 and p = 0.6) P [X 3] = 0.663 P [X = 2] = P [X 2] P [X 1] = 0.317 0.087 = 0.230 P [X > 0] = 1 P [X 0] = 1 0.010 = 0.90 Statistics packages are better than tables, as they don t limit the choice of n and p. Discrete Distributions 5

Discrete Distributions 6

The moments of the distribution are E[X] = np; V ar(x) = np(1 p); SD(X) = np(1 p) The value for E[X] can be derived a couple of ways. One is by the definition n ( n E[X] = i i i=0 n ( n 1 = n i = np = np = np i=1 n 1 j=0 n 1 j=0 ) p i (1 p) n i ( n 1 j ) pp i 1 (1 p) (n 1) (i 1) ) p j (1 p) (n 1) j P [Y = j] where Y Bin(n 1, p) Discrete Distributions 7

or we can use the fact that X = n i=1 Y i and E[X] = n E[Y i ] As mentioned earlier, E[Y i ] = p and there are n of them in the sum. i=1 The variance can be derived a couple of ways as well. One approach uses the fact that E[X 2 ] = np + n(n 1)p 2 which can be determined a number of ways. From this V ar(x) = E[X 2 ] (E[X]) 2 = np + n(n 1)p 2 (np) 2 = np np 2 = np(1 p) Discrete Distributions 8

Another approaches uses the fact that if Y 1, Y 2,..., Y n are independent RVs and X = i Y i n V ar(x) = V ar(y i ) i=1 A Ber(p) has variance p(1 p), thus this approach gives us the same answer. Note that it s variances that add, not standard deviations. SD(X + Y ) SD(X) + SD(Y ) Discrete Distributions 9

Geometric and Negative Binomial Distribution Geometric: This distribution is also based on Bernoulli trials. In this case, we repeat the trials until we get a success and record which trial we get this first success. The flip a coin until the first head example used earlier is an example of the geometric distribution The PMF for the geometric distribution (Geo(p)) is p(k) = p(1 p) k 1 ; k = 1, 2,... Discrete Distributions 10

Discrete Distributions 11

The moments of the distribution are E[X] = 1 p ; V ar(x) = 1 p p 2 = 1 p ( ) 1 p 1 ; SD(X) = 1 p p The mean suggests that if the chance of a success is 1 in n, on average we need to wait n trials until we see our first success. For example, if we bought one ticket a week in the Mass Millions lottery 1 (p = 13,983,816 ) we expect to wait 13,983,816 weeks (268,919 years) before we win the grand prize for the first time. Discrete Distributions 12

For the smallest prize (3 out of 6), p = 0.0177, so we expect to wait 1 0.0177 = 56.66 weeks until our first win of this type. As the standard deviation for this case is 55.99, the actual time we need to wait could be much higher. There is a probability of 0.027 that the first 3 out of 6 win will take over 200 weeks. Discrete Distributions 13

Negative Binomial: This is similar to the geometric distribution except it is the time for the rth success. The PMF for the negative binomial distribution (N Bin(r, p)) is p(k) = ( ) k 1 p r (1 p) k r ; k = r, r + 1, r + 2,... r 1 Note that Geo(p) = NBin(1, p). Also NBin(r, p) can be thought of as the sum of r independent Geo(p) RVs. Note that some books and software define the geometric and negative binomial slightly differently. Instead of the waiting time until success r, they define it as the number of failures before success r. For example, the package R uses this alternate definition for its routines. Discrete Distributions 14

Discrete Distributions 15

The moments of the distribution are E[X] = r r(1 p) ; V ar(x) = p p 2 = r p ( ) 1 p 1 ; SD(X) = r(1 p) p These results come from the fact that the negative binomial is the sum of r geometric RVs. Discrete Distributions 16

Poisson This is a distribution that is used for counts, often of rare events. It was originally derived as a limiting distribution of the binomial by Poisson (see the argument in section 2.1.5). However it is useful for many situations. One example mentioned before was the number of alpha particles emitted from a radioactive source during a fixed time period. Other examples: Queuing theory. How many customers will enter a bank in the next 30 minutes. Wave damage to cargo ships Number of matings per year in a population of African elephants Discrete Distributions 17

The PMF for the Poisson distribution (P ois(λ)) is p(k) = λk k! e λ ; k = 0, 1, 2,... The parameter of the Poisson distribution is the mean. Note that often when the Poisson distribution is used, the response is the number of counts during some time period. So lambda is often given as λ = rate time So doubling the time period observed will double λ Discrete Distributions 18

Discrete Distributions 19

The moments of the distribution are E[X] = λ; V ar(x) = λ; SD(X) = λ E[X] = k=0 k λk k! e λ E[X(X 1)] = k=0 k(k 1)λk k! e λ = λ k=1 λk 1 (k 1)! e λ = λ j=0 λj j! e λ = λ = λ 2 = λ 2 k=2 λk 2 (k 2)! e λ E[X 2 ] = E[X(X 1)] + E[X] = λ 2 + λ which gives V ar(x) = E[X 2 ] (E[X]) 2 = λ Discrete Distributions 20

Example: Bomb hits in London during WWII (R.D. Clark, 1946) London was divided into N = 576 small blocks of 0.25(km) 2 Let λ be the average number of hits in a block by bombs. The estimate of this is ˆλ = 0.9323. For any such block, the X be the number of hits. (approximately). X P ois(λ) Compare with actual data (N k = # blocks with k hits) k N k N P [X = k] 0 229 226.74 1 211 211.39 2 93 98.54 3 35 30.62 4 7 7.14 5+ 1 1.57 Discrete Distributions 21

Hypergeometric The binomial distribution occurs when you sample from a population consisting of two types of objects ( Success & Failure ) with replacement. Most sampling however is done without replacement. The Hypergeometric distribution describes this situation. Assume that your population contains n members, r of which are Successes and n r of them are Failures. You want to draw m items from the population. Let X be the number of successes in the m draws. The PMF for the hypergeometric distribution is p(k) = ( r n r ) k)( m k ( n ; k = 0, 1, 2,..., m m) Discrete Distributions 22

An example of the hypergeometric distribution was the Mass Millions lottery. The probability of the the number of matches is hypergeometric with n = 49, r = 6, and m = 6 k P [X = k] 0 0.43596 1 0.41302 2 0.13238 3 0.01765 4 0.00097 5 0.000018 6 1/13,983,816 When n, r, and n r get big relative to m, the hypergeometric looks a lot like the Bin(m, p = r n ) distribution. In this situation, the probability of a success doesn t change much as members of the population are sampled. For the above example however the binomial approximation is poor. Discrete Distributions 23