Discrete Distributions Chapter 6

Similar documents
Discrete Distributions

STAT/MATH 395 A - PROBABILITY II UW Winter Quarter Moment functions. x r p X (x) (1) E[X r ] = x r f X (x) dx (2) (x E[X]) r p X (x) (3)

Mathematical Statistics 1 Math A 6330

Things to remember when learning probability distributions:

Binomial and Poisson Probability Distributions

RMSC 2001 Introduction to Risk Management

MFM Practitioner Module: Quantitative Risk Management. John Dodson. September 23, 2015

Discrete Distributions

Part (A): Review of Probability [Statistics I revision]

Continuous RVs. 1. Suppose a random variable X has the following probability density function: π, zero otherwise. f ( x ) = sin x, 0 < x < 2

p. 4-1 Random Variables

Lectures on Elementary Probability. William G. Faris

RMSC 2001 Introduction to Risk Management

Mixture distributions in Exams MLC/3L and C/4

Method of Moments. which we usually denote by X or sometimes by X n to emphasize that there are n observations.

Chapter 5. Chapter 5 sections

STAT/MATH 395 PROBABILITY II

Creating New Distributions

Statistics for Economists. Lectures 3 & 4

Math/Stat 352 Lecture 8

3 Continuous Random Variables

Chapter 4. Repeated Trials. 4.1 Introduction. 4.2 Bernoulli Trials

Relationship between probability set function and random variable - 2 -

Department of Mathematics

Chapters 3.2 Discrete distributions

Lecture 6. Probability events. Definition 1. The sample space, S, of a. probability experiment is the collection of all

1 Review of Probability


1 Presessional Probability

Department of Mathematics

Chapter 3. Discrete Random Variables and Their Probability Distributions

Introduction and Overview STAT 421, SP Course Instructor

Distribution of a Sum of Random Variables when the Sample Size is a Poisson Distribution

M(t) = 1 t. (1 t), 6 M (0) = 20 P (95. X i 110) i=1

Topic 3: The Expectation of a Random Variable

or E ( U(X) ) e zx = e ux e ivx = e ux( cos(vx) + i sin(vx) ), B X := { u R : M X (u) < } (4)

PANJER CLASS UNITED One formula for the probabilities of the Poisson, Binomial, and Negative Binomial distribution.

ARCONES MANUAL FOR THE SOA EXAM P/CAS EXAM 1, PROBABILITY, SPRING 2010 EDITION.

1 Inverse Transform Method and some alternative algorithms

15 Discrete Distributions

Lecture 17: The Exponential and Some Related Distributions

Moments. Raw moment: February 25, 2014 Normalized / Standardized moment:

Definition: A random variable X is a real valued function that maps a sample space S into the space of real numbers R. X : S R

STAT 430/510 Probability Lecture 12: Central Limit Theorem and Exponential Distribution

STAT 302 Introduction to Probability Learning Outcomes. Textbook: A First Course in Probability by Sheldon Ross, 8 th ed.

Review 1: STAT Mark Carpenter, Ph.D. Professor of Statistics Department of Mathematics and Statistics. August 25, 2015

Exam 3, Math Fall 2016 October 19, 2016

Lecture 13. Poisson Distribution. Text: A Course in Probability by Weiss 5.5. STAT 225 Introduction to Probability Models February 16, 2014

STAT 3610: Review of Probability Distributions

Probability and Statistics Notes

Suppose that you have three coins. Coin A is fair, coin B shows heads with probability 0.6 and coin C shows heads with probability 0.8.

Week 1 Quantitative Analysis of Financial Markets Distributions A

Computer Science, Informatik 4 Communication and Distributed Systems. Simulation. Discrete-Event System Simulation. Dr.

1 Bernoulli Distribution: Single Coin Flip

Chapter 8. Some Approximations to Probability Distributions: Limit Theorems

Monotonicity and Aging Properties of Random Sums

n! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2

Limiting Distributions

Lecture 3. Discrete Random Variables

Basic concepts of probability theory

n! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2

Errata for the ASM Study Manual for Exam P, Fourth Edition By Dr. Krzysztof M. Ostaszewski, FSA, CFA, MAAA

Three hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER.

Poisson approximations

Random variables and transform methods

12d. Regular Singular Points

Probability Distributions Columns (a) through (d)

System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models

Brief Review of Probability

Conditional distributions (discrete case)

7.3 Singular points and the method of Frobenius

Estimates for probabilities of independent events and infinite series

STAT2201. Analysis of Engineering & Scientific Data. Unit 3

Limiting Distributions

STAT 414: Introduction to Probability Theory

Solutions to the Spring 2015 CAS Exam ST

ASM Study Manual for Exam P, First Edition By Dr. Krzysztof M. Ostaszewski, FSA, CFA, MAAA Errata

Continuous Probability Spaces

Discrete Probability Distributions

Equations with regular-singular points (Sect. 5.5).

Mathematical Methods for Computer Science

Basic concepts of probability theory

Module 3. Function of a Random Variable and its distribution

6. Bernoulli Trials and the Poisson Process

ASM Study Manual for Exam P, Second Edition By Dr. Krzysztof M. Ostaszewski, FSA, CFA, MAAA Errata

ECE 313 Probability with Engineering Applications Fall 2000

STA 4321/5325 Solution to Extra Homework 1 February 8, 2017

Basic concepts of probability theory

Stat 516, Homework 1

Lecture 10. Failure Probabilities and Safety Indexes

Actuarial Science Exam 1/P

Probability and Distributions

14 Branching processes

Expectations. Definition Let X be a discrete rv with set of possible values D and pmf p(x). The expected value or mean value of X, denoted by E(X ) or

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

CAS Exam MAS-1. Howard Mahler. Stochastic Models

Practice Exam 1. (A) (B) (C) (D) (E) You are given the following data on loss sizes:

Stat 134 Fall 2011: Notes on generating functions

Modelling the risk process

Solutions to Problem Set 4

Transcription:

Discrete Distributions Chapter 6 Negative Binomial Distribution section 6.3 Consider k r, r +,... independent Bernoulli trials with probability of success in one trial being p. Let the random variable X be the trial number k needed to have r-th success. Equivalently in the first k trials there are r successes (no matter when they have occurred) and the k-th trial must be a success. Since the trials are independent, the required probability is found by multiplying the two sub-probabilities: ( ) k p r ( p) k r p r ( ) k p r ( p) k r r where here we are assuming that k r, r +,.... Putting n k r, or equivalently k n + r, we can write the above equality in terms of n : ( ) n + r P (Y n) p r ( p) n n 0,, 2,... where Y X r r But we know from Combinatorics that ( ) ( n+r r n+r ) n, therefore we have ( ) n + r P (Y n) p r ( p) n n 0,, 2,... n Finally by a change of variable p ( n + r P (Y n) n, we can write: ) ( + ) r ( ) n n 0,, 2,... + Recall that therefore: ( ) n + r n (n + r )(n + r 2) (r) n! P (Y n) (n + r )(n + r 2) (r) n! ( ) r ( ) n n 0,, 2,... + +

In this new form, r can be taken to be any positive number (and not just a positive integer). So the negative binomial distribution has two positive parameters > 0 and r > 0. This distribution has an advantage over the Poisson distribution in modeling because it has one more parameter. To be able to use the table of the textbook directly, let us change n to k thoroughly: P (Y k) (k + r )(k + r 2) (r) ( ) r ( ) k k 0,, 2,... + + P (Y k) k terms {}}{ (r)(r + ) (k + r ) k ( + ) k+r k 0,, 2,... We need the values (k+r )(k+r 2) (r) to be able to calculate these probabilities. But note that Γ(k+r) (k+r )(k+r 2) (r)γ(r) (k + r )(k + r 2) (r) Γ(k + r) Γ(k + ) Γ(r) and tables of the values the Gamma function or log-gamma functions are provided in programming languages, so you can calculate the values (k+r )(k+r 2) (r) through the formula Γ(k+r) Γ(k+) Γ(r). Here is the reason for choosing the name Negative Binomial For a moment let us recall the binomial series we have learned in Calculus: ( + z) α + k which is valid for all real number α and all < z <. As an example: 4 + z ( + z) 4 + k (α)(α ) (α k + ) z k ( 4 )( 5 4 ) ( 4 k + ) z k < z < 2

Now change z to z, and substitute r for α to get: ( z) r + ( r)( r ) ( r k+) k ( z) k + (r)(r+) (r+k ) k z k dropping ( ) k + (r+k )(r+k 2) (r) k z k rearranging The series expansion k0 ( z) r ( k+r ) r z k is called the negative binomial expansion. ( z) r k0 ( k + r r ) z k Now we calculate the PGF of the negative binomial distribution. ( ) ( ) k + r r ( P N (z) E(z N ) z k P (N k) z k r + + k0 k0 ( ) r ( ) ( ) k + r k ( ) r z k + r + + k0 k0 ( ) r ( z ) r ( ) r ( + ( z) + + + + ( ) r ( + + + ( z) ) r ) r ( k + r r ( + ( z)) r ( (z )) r ) k ) ( z + ) k Expected value and Variance p(z) ( + ( z)) r p (z) r( + ( z)) r p (z) r(r + ) 2 ( + ( z)) r 2 3

p () r p () r(r + ) 2 E(N) p () r Var(N) p () + p ()( p ()) r + r 2 r( + ) Note. As we see, in the negative binomial distribution, the variance is larger than the expected value while in the Poisson distribution they are equal, therefore in modeling the data in which the sample variance seems to be larger than the sample mean, the negative binomial distribution is preferred over the Poisson distribution. If P X (z) is the PGF of X, then E(X) p () Var(X) p () + p ()( p ()) Theorem (Poisson as a limit of NBinomials). Let X n NBinomial(r n, n ) such that r n and n 0 and r n n λ > 0. Then X n d Poisson(λ) Note. Before proving this theorem, we recall from Calculus that when x 0, then the functions ln( + x) and x are equivalent in the sense that lim x 0 ln(+x) x. To see this equality apply the L Hospital rule. Once we have this, then in the quotients we can substitute x for ln( + x) whenever x 0. In fact, if g(x) is any function of x, then ln( + x) lim lim x g(x) x ln( + x) x ( x g(x) lim x ) ( ln( + x) lim x x ) x lim g(x) x x g(x) Proof of the Theorem. Set λ n r n n. Then from the assumption we have λ n λ. 4

Further: ( ) rn lim n P Xn (z) lim + n ( z) [ ] lim exp r n ln( + n ( z)) [ { }] exp lim r n ln( + n ( z)) [ { exp lim [ exp lim }] ln( n( z)) rn { ln(+ λ n( z) rn rn ) }] [ { exp lim λn( z) rn rn }] [ { }] exp lim λ n ( z) exp(λ(z )) So, we have proved that the PGF of the sequence X n s tends to the PGF of the Poisson(λ) distribution. Whence the claim. 5

Geometric Distribution The Geometric random variable with parameter 0 < q < is a variable with support {, 2,...} such that X k is the event that in a series of Bernoulli trials the first success occurs at time k. Since the first k trials should result in a failure, then P (X k) q ( q) k k, 2,... Note that P (X m) {q( q) m + q( q) m + q( q) m+ + } q( q) m { } + ( q) + ( q) 2 + q( q) m ( q)m ( q) Then for k, 2,... we have : P (X n + k X n) P (X n + k, X n) P (X n) P (X n + k) P (X n) ( q)n+k ( q) n ( q) k This property is called the memoryless property. Given that there are at least n claims, the probability distribution of the number of claims in excess of n does not depend on n : P (X n + k X n) P (X n + k X n) P (X n + k + X n) ( q) k ( q) k+ does not depend on n A random variable X has memoryless distribution if for all x s the conditional distribution (X X x) is the same for all x s. We may consider the Geometric distribution as a special case of Negative Binomial : In fact by changing k to k + and substituting Geometric() P (N k) for p we can write the probabilities in the new form: ( ) ( ) k + + k ( + ) k+ k 0,,... So, a Geometric distribution is a Negative Binomial distribution with r. Note that in this 6

new shape, we have P (N 0) + i.e. the value of the probability function at k 0 equals the probability of success. 7

(a,b,0) class section 6.5 Definition. Let p k P (X k), k 0,, 2,... be the probabilities of a discrete random variable. If there are two numbers a and b satisfying p k p k a + b k equivalently k p k p k ak + b k, 2,... then we say that X belongs to the class (a, b, 0). Note that the following four distributions are in class (a, b, 0) : Distribution a b p 0 Binomial(m, q) q q (m + )a ( q) m Negative Binomial(r, ) Geometric(q) (r )a ( + ) r 0 ( + ) Poisson(λ) 0 λ e λ As it has been shown in the literature, these four distributions are the only non-trivial distributions of the class (a, b, 0) Example. Consider a discrete random variable N of class of the (a, b, 0). Assume that P (X 0) 0., P (X ) 0.3, and P (X 2) 0.3 Calculate P (X 3). Solution. p k ( a + b ) p k k ( ) Putting k in ( ) we get: p (a + b)p 0 0.3 (a + b)(0.) a + b 3 Putting k 2 in ( ) we get: 8

p 2 (a + 2 b)p 0.3 (a + 2 b)(0.3) a + 2 b By solving the two equations thus found, we will have: a and b 4. Then the equality (8) reduces to Put k 3 in ( ) to have: p k ( + 4 ) p k k ( ) p 3 ( + 4 3 )p 2 p 3 3 p 2 3 (0.3) 0.0 Example. The Independent Insurance Company insures 25 risks, each with a 4% probability of loss. The probabilities of loss are independent. On average, how often would 4 or more risks have losses in the same year? A. Once in 3 years B. Once in 7 years C. Once in 39 years D. Once in 60 years E. Once in 72 years Solution. The distribution of loss is Binomial(m 25, q 0.04). Either find the probabilities through elementary formula p k m(m ) (m k+) q k ( q) m k or use the recursive formula: a q q 0.047 b (m + )a.08333 p 0 ( q) m (0.96) 25 0.3604 p (a + b)p 0 0.3754 p 2 (a + b 2 )p 0.877 p 3 (a + b 3 )p 2 0.0600 desired probability (p 0 + p + p 2 + p 3 ) 0.9835 0.065 4 or more risks have losses in the same year on average once in: /.065 60.6 years. Example. The distribution of accidents for 84 randomly selected policies is as follows: 9

Number of accidents Number of Policies 0 32 26 2 2 3 7 4 4 5 2 6 Total 84 Which of the following models best represents these data? (A) (B) (C) (D) (E) Negative binomial Discrete uniform Poisson Binomial Either Poisson or Binomial Solution. Number of accidents Number of Policies k p k p k 0 32 26 0.8 2 2 0.92 3 7.75 4 4 2.29 5 2 2.50 6 3.00 Total 84 0

Since the values k p k p k increase, a negative binomial distribution is best fit. Example. A discrete probability distribution has the following properties: (i) p k c( + /k)p k for k, 2,... (ii) p 0 0.5. Calculate c. Solution. This distribution is of class (a, b, 0) with a b c. Since the probabilities are non-negative we must have c 0. A value of zero for c result in p p 2 0 which together with p 0 0.5 do not result in the sum of probabilities being equal to. So we have c > 0 and then a > 0. So the distribution is of negative binomial. For the negative binomial we have a and b (r ). Then a b r r 2 p 0 0.5 0.5 ( + ) r ( + ) 2 2 0.442 c a + 0.442.442 0.293 Example. You are given the following: A portfolio consists of 0 independent risks. The distribution of the annual number of claims for each risk in the portfolio is given by a Poisson distribution with mean λ 0.. Determine the probability of the portfolio having more than claim per year. A. 5% B. 0% C. 26% D. 37% E. 63% Solution. Since the sum of independent Poisson r.v. s is distributed as Poisson, this portfolio has a Poisson distribution with parameter λ (0)(0.). So the probability of more than

one claim in one year is ( (p 0 + p ) e + e ) 0.264 2

Truncation and modification at zero section 6.6 The discrete distributions in the class (a, b, 0) all have positive probability at zero p(0) > 0. But if in a study the number of claims is under study for onlt those losses that have resulted in a claim, then the minimum observed value is and therefore p(0) 0. We can create such a counting distribution from a class (a, b, 0) by assigning the zero value to p(0) and dividing the remaining probabilities by p(0). Such a new distribution is called a zero-truncated distribution. In some insurance situations the chance of no claim is high so we may have a counting data that needs a large value of p(0) to be properly modeled. Therefore we may modify the counting model by assigning a large value to p(0) and then modify the other values of p(k) so as to have a probability distribution. This new distribution is called zero-modified. Note that in both cases of zero-truncated distribution and zero-modified distribution, the old probabilities {p(), P (2),...} are all multiplied by a constant, therefore the ratio remains the same as before for k 2, 3,..., p(k) p(k ) p(k) p(k ) a + b k k 2, 3,... A distribution with this property is said to of class (a, b, ). So by modifying a (a, b, 0) class member we get a (a, b, ) class member. Definition. A discrete random variable X {0,, 2,...} is said to belong to class (a, b, ) if its density function satisfies the equality p(k) ( a + b ) p(k ) k 2, 3,... k Note. The class (a, b, ) satisfies the same recursive form as the class (a, b, 0) does, but in the case of (a, b, ) the value p(0) is not as part of recursion. 3

Let us denote the probability function of (a, b, 0) by p(k) and that of (a, b, ) by p M (k). By assigning any number from γ [0, ) to p M (0), we then define p M by (zero-modified distribution) p M γ k 0 (k) γ p(0) p(k) k, 2,... We here verify that this is indeed a probability function: p M (0) + p M γ (k) γ + p(0) p(k) γ + γ p(0) k k γ + γ ( p(0)) γ + ( γ) p(0) The special case of γ 0 : (zero-truncated distribution) p T (k) p(k) k 0 k 0 p(0) p(k) k, 2,... Example. Let X be distributed as Geometric( 9). Calculate (i) The zero-truncated distribution. (ii) The zero-modified distribution with P M (0) 0.3 Solution. Part (i): p(0) + 0 p(0) 0.9 0 9 k 9k ( + ) k+ 0 k+ p T 0 k 0 (k) p(0) p(k) 0 9 9 k 9k 0k+ 0 k 9 k 0 k k, 2,... (zero-truncated distribution) 4

Part (ii): γ p(0) 0.7 0.9 7 9 γ p(0) p(k) 7 9 0 k+ 7(9k ) 0 k+ p M 0.3 k 0 (k) 7(9 k ) k, 2,... (zero-modified distribution) 0 k+ 9 k In the following discussion, for the sake of simplicity, we denote the base probabilities by {p 0, p,...} and the resulted probabilities due to truncation or modification by {q 0, q,...}. Theorem. Let N be in the class (a, b, 0) having the MGF M(t) and the PGF Q(z). Then the MGF and PGF of the the modified (and truncated) random variable are equal to ( ) M (t) γ p 0 p 0 + γ p 0 M(t) a weighted average ( ) Q (t) γ p 0 p 0 + γ p 0 Q(z) a weighted average 5

Proof. We only prove the first equality as the proof of the second one is similar. M (t) k0 ekt p M (k) γ + γ p 0 k ekt p k γ + γ p 0 { k0 ekt p k p 0 } { } γ + γ p 0 M(t) p 0 γ p 0 γ p 0 + γ p 0 M(t) γ p 0 p 0 + γ p 0 M(t) Note. The constant function is the MGF of the degenerate random variable which give the probability to k 0. So, the above equalities show that the modified distribution is the mixture of this degenerate distribution and the base distribution. 6

Extended Truncated Negative Binomial and Logarithmic Distributions Definition. By the extended truncated negative binomial distribution we mean the discrete distribution whose probability function is of the form: p k ( a + b ) k pk k 2, 3,... p 0 0 where a, b (r ), r 0, r >, > 0 Note. The requirements r > and > 0 are needed to ensure that this indeed defines a probability function as the following lemma shows. Lemma. ( ) k (r + )(r + 2) (r + k ) p k p + k 2, 3,... Proof. From the defining recursive formula, we have ( (r ) p k + + k( + ) Writing this for k 2, 3,... gives us: Multiplying these gives the result. ) k + (r ) (r + k ) p k p k p k k( + ) k( + ) p 2 p 3. p k r+ 2 p r+2 3 p 2 r+k k p k + r + k k p k Lemma. For any choice of 0 < p we have k p k <. Proof. Equivalently we must show that the series with positive terms ( ) k (r + )(r + 2) (r + k ) p k p + p + k k2 7

is convergent. We use the so-called ratio-test to achieve this. Set ( ) k (r + )(r + 2) (r + k ) a k + Then lim a k+ a k lim k k + r + k k + so, according to the ratio-test, the series is convergent. + < Note. Here is a closed form for p : k p ( ) k k p + p (r+)(r+2) (r+k ) k2 ( ) p + p ( ) k ( r)( r )( r 2) ( r k+) r k2 ( ) {[ p + p r + ( ) ] } k ( r)( r )( r 2) ( r k+) k r ( ) { ( ) } r p + p r r ( ) { } p + p r ( + ) r r ( ) { } p + p () r+ (+r) r { } p + p () r+ (+r) r { } p + ()r+ (+r) r { } p () r+ r Therefore p r () r+ 8

and for this one needs the assumption r 0 otherwise the denominator would be zero. Note. We recall from Calculus that d dr ( + )r+ ( + ) r+ ln( + ) Then by applying the L Hospital rule we will have: lim p lim r 0 r 0 So then for k 2 we have r ( + ) r+ lim r 0 lim r 0 p k lim r 0 p ( ( + ) r+ ln( + ) ( + ) ln( + ) ) k (r+)(r+2) (r+k ) ( ) k () (k ) () ln() ( ) k ln() k This and lim r 0 p () ln() lim p k k can be put together : ( + ) k ln( + ) k k, 2,... But the values on the right-hand side are positive numbers that sum to : ( ) k ( ) k ( ) ( ( )) k ln() k ln() k k ln() ln ( ln() ) ( ( )) ( ln ln() ) ln( + ) definition. The distribution with probability function ( ) k p(k) + ln( + ) k is called Logarithmic distribution Note. The class (a, b, ) has 4 members. The list of members is given on page 94 of the textbook and the distributional information about these distributions is given at the end of the textbook. 9