Statistics for Engineers Lecture 2 Discrete Distributions

Similar documents
Chapter 3: Discrete Random Variable

STAT509: Discrete Random Variable

STAT 509 STATISTICS FOR ENGINEERS

Statistics for Economists. Lectures 3 & 4

Chapter 2: Discrete Distributions. 2.1 Random Variables of the Discrete Type

Chapter 3. Discrete Random Variables and Their Probability Distributions

Lecture 3. Discrete Random Variables

STAT 509 STATISTICS FOR ENGINEERS. Lecture Notes. Dewei Wang. Department of Statistics. University of South Carolina

Mathematical Statistics 1 Math A 6330

Lecture 13. Poisson Distribution. Text: A Course in Probability by Weiss 5.5. STAT 225 Introduction to Probability Models February 16, 2014

Chapter 3 Discrete Random Variables

Chapter 3. Discrete Random Variables and Their Probability Distributions

An-Najah National University Faculty of Engineering Industrial Engineering Department. Course : Quantitative Methods (65211)

II. The Binomial Distribution

BINOMIAL DISTRIBUTION

Topic 3 - Discrete distributions

b. ( ) ( ) ( ) ( ) ( ) 5. Independence: Two events (A & B) are independent if one of the conditions listed below is satisfied; ( ) ( ) ( )

Common Discrete Distributions

Known probability distributions

Relationship between probability set function and random variable - 2 -

Math 151. Rumbos Fall Solutions to Review Problems for Final Exam

Practice Exam 1 (Sep. 21, 2015)

Discrete Distributions

Topic 3: The Expectation of a Random Variable

Sample Problems for the Final Exam

Week 6, 9/24/12-9/28/12, Notes: Bernoulli, Binomial, Hypergeometric, and Poisson Random Variables

Fault-Tolerant Computer System Design ECE 60872/CS 590. Topic 2: Discrete Distributions

Binomial random variable

Chapter 3: Discrete Random Variables

Notes for Math 324, Part 17

Brief Review of Probability

Discrete Distributions

Random Variables Example:

N = total number of individuals = 8 r = number who had internships ( Class 1 ) = 5 N r = number who had student teaching ( Class 2 ) = 3.

Introduction to Probability and Statistics Slides 3 Chapter 3

Discrete random variables and probability distributions

ELEG 3143 Probability & Stochastic Process Ch. 2 Discrete Random Variables

Random Variable And Probability Distribution. Is defined as a real valued function defined on the sample space S. We denote it as X, Y, Z,

2.3 Analysis of Categorical Data

Introduction to Statistical Data Analysis Lecture 3: Probability Distributions

Binomial and Poisson Probability Distributions

p. 4-1 Random Variables

Probability and Statistics for Engineers

Contents 1. Contents

MATH 250 / SPRING 2011 SAMPLE QUESTIONS / SET 3

(Ch 3.4.1, 3.4.2, 4.1, 4.2, 4.3)

Bernoulli and Binomial Distributions. Notes. Bernoulli Trials. Bernoulli/Binomial Random Variables Bernoulli and Binomial Distributions.

Basic concepts of probability theory

Errata for the ASM Study Manual for Exam P, Fourth Edition By Dr. Krzysztof M. Ostaszewski, FSA, CFA, MAAA

Basic concepts of probability theory

Learning Objectives for Stat 225

Statistics for Engineers Lecture 4 Reliability and Lifetime Distributions

Statistics for Economists Lectures 6 & 7. Asrat Temesgen Stockholm University

3.4. The Binomial Probability Distribution

1 Inverse Transform Method and some alternative algorithms

Statistical Quality Control IE 3255 Spring 2005 Solution HomeWork #2

Discrete Random Variables (cont.) Discrete Distributions the Geometric pmf

Closed book and notes. 120 minutes. Cover page, five pages of exam. No calculators.

Northwestern University Department of Electrical Engineering and Computer Science

ASM Study Manual for Exam P, Second Edition By Dr. Krzysztof M. Ostaszewski, FSA, CFA, MAAA Errata

Probability theory and inference statistics! Dr. Paola Grosso! SNE research group!! (preferred!)!!

Bernoulli Trials, Binomial and Cumulative Distributions

(Ch 3.4.1, 3.4.2, 4.1, 4.2, 4.3)

Suppose that you have three coins. Coin A is fair, coin B shows heads with probability 0.6 and coin C shows heads with probability 0.8.

Topic 3: The Expectation of a Random Variable

ECE-580-DOE : Statistical Process Control and Design of Experiments Steve Brainerd 27 Distributions:

Discrete Distributions

Geometric Distribution The characteristics of a geometric experiment are: 1. There are one or more Bernoulli trials with all failures except the last

15 Discrete Distributions

Stat Lecture 20. Last class we introduced the covariance and correlation between two jointly distributed random variables.

System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models

1. Summarize the sample categorical data by creating a frequency table and bar graph. Y Y N Y N N Y Y Y N Y N N N Y Y Y N Y Y

Discrete Random Variable

System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models

Question Points Score Total: 137

POISSON RANDOM VARIABLES

STA 4321/5325 Solution to Extra Homework 1 February 8, 2017

Discrete Random Variables

2014 SM4 Revision Questions Distributions

Exponential, Gamma and Normal Distribuions

Random Variables. Definition: A random variable (r.v.) X on the probability space (Ω, F, P) is a mapping

Math 151. Rumbos Fall Solutions to Review Problems for Final Exam

Computer Science, Informatik 4 Communication and Distributed Systems. Simulation. Discrete-Event System Simulation. Dr.

A) Questions on Estimation

DISCRETE VARIABLE PROBLEMS ONLY

STA 584 Supplementary Examples (not to be graded) Fall, 2003

Continuous-Valued Probability Review

HW Solution 4 Not Due

1. If X has density. cx 3 e x ), 0 x < 0, otherwise. Find the value of c that makes f a probability density. f(x) =

The Bernoulli distribution has only two outcomes, Y, success or failure, with values one or zero. The probability of success is p.

Discrete Distributions

Notes 12 Autumn 2005

Probability and Stochastic Processes: A Friendly Introduction for Electrical and Computer Engineers Roy D. Yates and David J.

2. The Standard Normal Distribution can be described as a. N(0,1) b.n(1,0)

Lecture 6. Probability events. Definition 1. The sample space, S, of a. probability experiment is the collection of all

STAT Examples Based on all chapters and sections

Lecture 14. Text: A Course in Probability by Weiss 5.6. STAT 225 Introduction to Probability Models February 23, Whitney Huang Purdue University

Probability Distributions

Basic concepts of probability theory

Transcription:

Statistics for Engineers Lecture 2 Discrete Distributions Chong Ma Department of Statistics University of South Carolina chongm@email.sc.edu January 18, 2017 Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 1 / 37

Outline 1 Discrete Distribution 2 Binomial Distribution 3 Geometric Distribution 4 Negative Binomial Distribution 5 Hypergeometric Distribution 6 Poisson Distribution Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 2 / 37

Discrete Distribution Suppose that Y is a discrete random variable. The function P Y (y) = P(Y = y) is called the probability mass function(pmf) for Y. The pmf p Y (y) is a function that assigns probabilities to each possible value of Y, satisfying the following 1 0 < p Y (y) < 1, for all possible values of y. 2 The sum of the probabilities, taken over all possible values of Y, must equal 1; i.e., y p Y (y) = 1. The cumulative distribution function(cdf) of Y is F Y (y) = P(Y y) 1 The cfd F Y (y) is a nondecreasing function. 2 0 F Y (y) 1 Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 3 / 37

Discrete Distribution The expected value of Y is given by µ = E(Y ) = y yp Y (y) The variance of Y is given by σ 2 = var(y ) = E[(Y µ) 2 ] = y (y µ) 2 p Y (y) The standard deviation of Y is given by σ = σ 2 = var(y ) Equivalently, var(y ) = E(Y 2 ) [E(Y )] 2. The expected value for a discrete random variable Y is a weighted average of the possible values of Y. The variance is weighted distance(squared difference) of the possible values of Y from the mean. Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 4 / 37

Discrete Distribution Let Y be a discrete r.v. with pmf p Y (y). Suppose that g is a real-valued function. Then g(y ) is a random variable and E[g(Y )] = y g(y)p Y (y) Furthermore, let g 1, g 2,..., g k are real-valued functions, and c is any real constant. Expectations satisfy the following (linearity) properties: E(c) = c E(cg(Y )) = ce(g(y )) E( k j=1 g j(y )) = k j=1 E[g j(y )] Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 5 / 37

Discrete Distribution Example A mail-order computer business has six telephone lines. Let Y denote the number of lines in use at a specific time. Suppose that the probability mass function(pmf) of Y is given by y 0 1 2 3 4 5 6 p Y (y) 0.10 0.15 0.20 0.25 0.20 0.06 0.04 The expected value of Y is µ = E(Y ) = y yp Y (y) = 0(0.10) + 1(0.15) + 2(0.20) + 3(0.25) + 4(0.20) + 5(0.06) + 6(0.04) = 2.64 Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 6 / 37

Discrete Distribution The variance of Y is σ 2 = E[(Y µ) 2 ] = y (y µ) 2 P Y (y) = (0 2.64) 2 0.10 + (1 2.64) 2 0.15 + (2 2.64) 2 0.20 + (3 2.64) 2 0.25 + (4 2.64) 2 0.20 + (5 2.64) 2 0.06 + (6 2.64) 2 0.04 = 2.37 Alternatively, Note that E(Y 2 ) = y y 2 P Y (y) = 0 2 (0.10) + 1 2 (0.15) + 2 2 (0.20) + 3 2 (0.25) + 4 2 (0.20) + 5 2 (0.06) + 6 2 (0.04) = 9.34 Thus, σ 2 = E(Y 2 ) [E(Y )] 2 = 9.34 2.64 2 = 2.37 Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 7 / 37

Discrete Distribution (a) What is the probability that exactly two lines are in use? p Y (2) = P(Y = 2) = 0.20 (b) What is the probability that at most two lines are in use? P(Y 2) = F Y (2) = P(Y = 0) + P(Y = 1) + P(Y = 2) = p Y (0) + p Y (1) + p Y (2) = 0.10 + 0.15 + 0.20 = 0.45 (c) What is the probability that at least five lines are in use? P(Y 5) = F Y (5) = 1 F Y (5) = P(Y = 5) + P(Y = 6) = p Y (5) + p Y (6) = 0.06 + 0.04 = 0.10 Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 8 / 37

Outline 1 Discrete Distribution 2 Binomial Distribution 3 Geometric Distribution 4 Negative Binomial Distribution 5 Hypergeometric Distribution 6 Poisson Distribution Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 9 / 37

Binomial Distribution Bernoulli trials: Many experiments can be considered as consisting of a sequence of trials such that Each trial results in a success or a failure. The trials are independent. The probability of success, denoted by p, is the same on every trial. Examples 1 When circuit boards used in the manufacture of Blue Ray players are tested, the long-run percentage of defective boards is 5%. circuite board = trial defective board is observed = success p = P( success ) = P(defective board) = 0.05 2 Ninety-eight percent of all air traffic radar signals are correctly interpreted the first time they are transmitted. radar signal = trial signal is correctly interpreted = success p = P( success ) = P(correct interpretation) = 0.98 Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 10 / 37

Binomial Distribution Suppose that n Bernoulli trials are performed. Define Y = the number of successes(out of n trials performed) we say that Y has a binomial distribution with number of trials n and success probability p, denoted by Y b(n, p). The probability mass function(pmf) of Y is given by {( n ) y p y (1 p) n y, y = 0, 1,..., n p Y (y) = 0, otherwise The mean/variance of Y are µ = E(Y ) = np, σ 2 = var(y ) = np(1 p) Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 11 / 37

Binomial Distribution PMF 0.00 0.05 0.10 0.15 0.20 0.25 CDF 0.0 0.2 0.4 0.6 0.8 1.0 0 2 4 6 8 10 y 0 2 4 6 8 10 y Figure 1: Plots of PMF and CDF for Binomial distribution Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 12 / 37

Binomial Distribution Example In an agricultural study, it is determined that 40% of all plots respond to a certain treatment. Four plots are observed. In this situation, we interpret that plot of land = trial plot responds to treatment = success p=p( success )=P(responds to treatment)=0.4 If the Bernoulli trial assumptions hold(independent plots, same response probability for each plot), then Y = the number of plots which respond b(n = 4, p = 0.4) (a) What is the probability that exactly two plots respond? (b) What is the probability that at least one plot responds? (c) what are E(Y ) and var(y )? Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 13 / 37

Binomial Distribution (a) What is the probability that exactly two plots respond? ( ) 4 P(Y = 2) = (0.4) 2 (1 0.4) 2 2 = 6(0.4) 2 (0.6) 2 = 0.3456 (b) What is the probability that at least one plot responds? P(Y 1) = 1 P(Y = 0) ( ) 4 = 1 (0.4) 0 (1 0.4) 4 0 = 1 (0.6) 4 = 0.8704 (c) what are E(Y ) and var(y )? E(Y ) = np = 4(0.4) = 1.6 var(y ) = np(1 p) = 4(0.4)(0.96) = 0.96 Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 14 / 37

Binomial Distribution Example An electronics manufacturer claims that 10% of its power supply units need servicing during the warranty period. Technicians at a testing laboratory purchase 30 units and simulate usage during the warranty period. We interpret power supply unit = trial supply unit needs servicing during warranty period = success p=p( success )=P(supply unit needs servicing)=0.1 (a) What is the probability that exactly five of the 30 power supply units require servicing during the warranty period? (b) What is the probability that at most five of the 30 power supply units require servicing during the warranty period? (c) What is the probability that at least five of the 30 power supply units require servicing during the warranty period? (d) What is P(2 Y 8)? Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 15 / 37

Binomial Distribution (a) What is the probability that exactly five of the 30 power supply units require servicing during the warranty period? p Y (5) = P(Y = 5) = ( ) 30 5 (0.1) 5 (0.9) 30 5 = 0.1023 (b) What is the probability that at most five of the 30 power supply units require servicing during the warranty period? F Y (5) = P(Y 5) = 5 ( 30 ) y=0 y (0.1) y (0.9) 30 y = 0.9268 (c) What is the probability that at least five of the 30 power supply units require servicing during the warranty period? P(Y 5) = 1 4 ( 30 ) (0.1) y (0.9) 30 y = 0.1755 y=0 (d) What is P(2 Y 8)? P(2 Y 8) = 8 y=2 y ( 30 y ) (0.1) y (0.9) 30 y = 0.8143 p Y (y) = P(Y = y) F Y (y) = P(Y y) dbinom(y,n,p) pbinom(y,n,p) Table 1: R code for Binomial Distribution Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 16 / 37

Outline 1 Discrete Distribution 2 Binomial Distribution 3 Geometric Distribution 4 Negative Binomial Distribution 5 Hypergeometric Distribution 6 Poisson Distribution Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 17 / 37

Geometric Distribution The geometric distribution also arises in experiments involving Bernoulli trials: 1 Each trial results in a success or a failure. 2 The trials are independent. 3 The probability of success, denoted by p, 0 < p < 1, is the same on each trial. Suppose that Bernoulli trials are continuously observed. Define Y = the number of trials to observe the first success We say that Y has a geometric distribution with success probability p. For short, Y geom(p). The probability mass function(pmf) of Y is { (1 p) y 1 p, y = 1, 2, 3,... p Y (y) = 0, otherwise Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 18 / 37

Geometric Distribution If Y geom(p), then mean E(Y ) = 1 p variance var(y ) = 1 p p 2 Example Biology students are checking the eye color of fruit flies. For each fly, the probability of observing while eyes is p = 0.25. We interpret friut fly = trial fly has while eyes = success p = P( success ) = 0.25 If the Bernoulli trial assumptions hold(independent flies, same probability of white eyes for each fly) Y = the number of flies needed to find the first white-eyed geom(p = 0.25) Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 19 / 37

Geometric Distribution (a) What is the probability the first white-eyed fly is observed on the fifth fly checked? p Y (5) = P(Y = 5) = (1 0.25) 5 1 (0.25) 0.079 (b) What is the probability the first white-eyed fly is observed before the fourth fly is examined? F Y (3) = P(Y 3) = P(Y = 1) + P(Y = 2) + P(Y = 3) = (1 0.25) 1 1 (0.25) + (1 0.25) 2 1 (0.25) + (1 0.25) 3 1 (0.25) = 0.25 + 0.1875 + 0.140625 0.578 p Y (y) = P(Y = y) F Y (y) = P(Y y) dgeom(y-1,p) pgeom(y-1,p) Table 2: R code for Geometric Distribution Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 20 / 37

Geometric Distribution PMF 0.00 0.05 0.10 0.15 0.20 0.25 CDF 0.0 0.2 0.4 0.6 0.8 1.0 0 5 10 15 20 25 30 y 0 5 10 20 30 y Figure 2: Plots of PMF and CDF for Geometric distribution Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 21 / 37

Outline 1 Discrete Distribution 2 Binomial Distribution 3 Geometric Distribution 4 Negative Binomial Distribution 5 Hypergeometric Distribution 6 Poisson Distribution Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 22 / 37

Negative Binomial Distribution The negative binomial distribution also arises in experiments involving Bernoulli trials: 1 Each trial results in a success or a failure. 2 The trials are independent. 3 The probability of success, denoted by p, 0 < p < 1, is the same on each trial. Suppose that Bernoulli trials are continuously observed. Define Y = the number of trials to observe the rth success We say that Y has a negative binomial distribution with success probability p. For short, Y nib(r, p). The probability mass function(pmf) of Y is {( y 1 p Y (y) = r 1) p r (1 p) y r, y = r, r + 1,... 0, otherwise Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 23 / 37

Negative Binomial Distribution If Y nib(r, p), then mean E(Y ) = r p variance var(y ) = r(1 p) p 2 Example At an automotive paint plant, 15% of all batches sent to the lab for chemical analysis do not conform to specifications. In this situation, We interpret batch = trial batch does not conform = success p = P( success ) = 0.15 If the Bernoulli trial assumptions hold(independent flies, same probability of white eyes for each fly) Y = the number of batches needed to find the third nonconforming nib(r = 3, p = 0.15) Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 24 / 37

Negative Binomial Distribution (a) What is the probability the third nonconforming batch is observed on the tenth batch sent to the lab? ( ) 10 1 p Y (10) = P(Y = 10) = (0.15) 3 (1 0.15) 10 3 3 1 ( ) 9 = (0.15) 2 (0.85) 7 0.039 2 (b) What is the probability no more than two nonconforming batches will be observed among the first 30 batches sent to the lab? Note: It implies the third nonconforming batch must be observed on the 31st batch tested, the 32nd, the 33rd, etc. P(Y 31) = 1 P(Y 30) 30 ( ) y 1 = 1 (0.15) 3 (1 0.15) y 3 0.151 3 1 y=3 Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 25 / 37

Negative Binomial Distribution PMF 0.00 0.01 0.02 0.03 0.04 0.05 CDF 0.0 0.2 0.4 0.6 0.8 1.0 10 30 50 70 y 0 20 40 60 y Figure 3: Plots of PMF and CDF for negative binomial distribution p Y (y) = P(Y = y) F Y (y) = P(Y y) dnbinom(y-r,r,p) pnbinom(y-r,r,p) Table 3: R code for negative binomial Distribution Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 26 / 37

Outline 1 Discrete Distribution 2 Binomial Distribution 3 Geometric Distribution 4 Negative Binomial Distribution 5 Hypergeometric Distribution 6 Poisson Distribution Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 27 / 37

Hypergeometric Distribution Setting: Consider a population of N objects and suppose that each object belongs to one of two dichotomous classes: class 1 and class 2. For example, the objects(classes) might be people(infected/not), parts(conforming/not), plots of land(respond to treatment/not), etc. In the population of interest, we have N = total number of objects r = number of objects in class 1 N r = number of objects in class 2 Envision taking a sample n objects from the population(objects are selected at random and without replacement). Define Y = the number of objects in class 1(out of the ne selected) We say that Y has a hypergeometric distribution, Y hyper(n, n, r). Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 28 / 37

Hypergeometric Distribution If Y hyper(n, n, r), the the probability mass function of Y is given by ( y)( r N r n y), y r and n y N r p Y (y) = ( N n) 0, otherwise The mean and variance of Y, mean E(Y ) = n( r N ) variance var(y ) = n( r N )( N r N N n )( N 1 ) p Y (y) = P(Y = y) F Y (y) = P(Y y) dhyper(y,r,n-r,n) phyper(y,r,n-r,n) Table 4: R code for hypergeometric Distribution Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 29 / 37

Hypergeometric Distribution Example A supplier ships parts to a company in lots of 100 parts. The company has an acceptance sampling plan which adopts the following acceptance rule:...sample 5 parts at random and without replacement. If there are no defectives in the sample, accept the entire lot; otherwise, reject the entire lot. The population size is N = 100, the sample size n = 5. The random variable Y = the number of defectives in the sample hyper(n = 100, n = 5, r) (a) If r=10, what is the probability that the lot will be accepted? (b) If r=10, what is the probability that at least 3 of the 5 parts sampled are defective? Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 30 / 37

Hypergeometric Distribution (a) If r=10, what is the probability that the lot will be accepted? p y (0) = ( 10 0 )( 90 5 ) ( 100 ) 0.584 5 (b) If r=10, what is the probability that at least 3 of the 5 parts sampled are defective? P(Y 3) = 1 P(Y 2) = 1 [p(y = 0) + P(Y = 1) + P(Y = 2)] [( 10 )( 90 ) ( 10 )( 90 ) ( 10 )( 90 )] 0 5 1 4 2 3 = 1 ( 100 ) + ( 100 ) + ( 100 ) 5 5 5 = 1 (0.584 + 0.339 + 0.070) 0.007 Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 31 / 37

Hypergeometric Distribution PMF 0.0 0.1 0.2 0.3 0.4 0.5 0.6 CDF 0.0 0.2 0.4 0.6 0.8 1.0 0 1 2 3 4 5 y 1 0 1 2 3 4 5 6 y Figure 4: Plots of PMF and CDF for Hypergeometric distribution Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 32 / 37

Outline 1 Discrete Distribution 2 Binomial Distribution 3 Geometric Distribution 4 Negative Binomial Distribution 5 Hypergeometric Distribution 6 Poisson Distribution Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 33 / 37

Possion Distribution The Poisson distribution is commonly used to model counts, such as 1 the number of customers entering a post office in a given hour. 2 the number of machine breakdowns per month 3 the number of insurance claims received per day 4 the number of defects on a piece of raw material In general, we define Y = the number of occurrence over a unit interval of time(or space) A Poisson distribution for Y emerges if occurrences obey the following postulates: P1. The number of occurrences in non-overlapping intervals are independent. P2. The probability of an occurrence is proportional to the length of the interval. P3. The probability of 2 or more occurrences in a sufficiently short interval is 0. Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 34 / 37

Possion Distribution We say that Y has a Poisson distribution, denoted by Y Poisson(λ). A process that produces occurrences according to these postulates is called a Poisson Process. If Y Poisson(λ), the probability mass function of Y is { λ y e λ y!, y = 0, 1, 2,... p Y (y) = 0, otherwise And the mean/variance of Y is mean E(Y ) = λ variance Var(Y ) = λ p Y (y) = P(Y = y) F Y (y) = P(Y y) dpois(y,λ ) ppois(y,λ) Table 5: R code for Poisson Distribution Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 35 / 37

Possion Distribution Example Let Y denote the number of times per month that a detectable amount of radioactive gas is recorded at a nuclear power plant. Suppose that Y follows a Poisson distribution with mean λ = 2.5 times per month. (a) What is the probability that there are exactly three times a detectable amount of gas is recorded in a given month? P(Y = 3) = (2.5)3 e 2.5 3! = 15.625e 2.5 6 0.214 (b) What is the probability that there are no more than three times a detectable amount of gas is recorded in a given month? P(Y 3) = P(Y = 0) + P(Y = 1) + P(Y = 2) + P(Y = 3) = (2.5)0 e 2.5 0! 0.544 + (2.5)1 e 2.5 1! + (2.5)2 e 2.5 2! + (2.5)3 e 2.5 3! Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 36 / 37

Possion Distribution PMF 0.00 0.05 0.10 0.15 0.20 0.25 CDF 0.0 0.2 0.4 0.6 0.8 1.0 0 2 4 6 8 10 y 0 2 4 6 8 10 y Figure 5: Plots of PMF and CDF for negative binomial distribution Chong Ma (Statistics, USC) STAT 509 Spring 2017 January 18, 2017 37 / 37