Mathematical Statistics 1 Math A 6330

Similar documents
Chapter 3 Discrete Random Variables

Statistics for Economists. Lectures 3 & 4

Chapters 3.2 Discrete distributions

Things to remember when learning probability distributions:

Chapter 5. Chapter 5 sections

Discrete Distributions

ELEG 3143 Probability & Stochastic Process Ch. 2 Discrete Random Variables

Lecture 13. Poisson Distribution. Text: A Course in Probability by Weiss 5.5. STAT 225 Introduction to Probability Models February 16, 2014

Discrete Distributions

Lecture 3. Discrete Random Variables

Relationship between probability set function and random variable - 2 -

Definition: A random variable X is a real valued function that maps a sample space S into the space of real numbers R. X : S R

Mathematical Statistics 1 Math A 6330

Chapter 6 Expectation and Conditional Expectation. Lectures Definition 6.1. Two random variables defined on a probability space are said to be

Continuous Probability Spaces

Discrete random variables and probability distributions

Chapter 3. Discrete Random Variables and Their Probability Distributions

More on Distribution Function

Chapter 2: Discrete Distributions. 2.1 Random Variables of the Discrete Type

SDS 321: Introduction to Probability and Statistics

Chapter 3. Discrete Random Variables and Their Probability Distributions

Topic 3: The Expectation of a Random Variable

Discrete Random Variables

Random Variables Example:

a zoo of (discrete) random variables

Stat 100a, Introduction to Probability.

Topic 3: The Expectation of a Random Variable

Week 2. Review of Probability, Random Variables and Univariate Distributions

Binomial and Poisson Probability Distributions

Chapter 2. Discrete Distributions

2. Suppose (X, Y ) is a pair of random variables uniformly distributed over the triangle with vertices (0, 0), (2, 0), (2, 1).

Introduction to Probability and Statistics Slides 3 Chapter 3

STAT/MATH 395 A - PROBABILITY II UW Winter Quarter Moment functions. x r p X (x) (1) E[X r ] = x r f X (x) dx (2) (x E[X]) r p X (x) (3)

Random Variables. Definition: A random variable (r.v.) X on the probability space (Ω, F, P) is a mapping

Northwestern University Department of Electrical Engineering and Computer Science

1 Review of Probability and Distributions

1 Bernoulli Distribution: Single Coin Flip

1 Random Variable: Topics

1 Presessional Probability

Lecture Notes 2 Random Variables. Discrete Random Variables: Probability mass function (pmf)

CDA6530: Performance Models of Computers and Networks. Chapter 2: Review of Practical Random Variables

Math/Stat 352 Lecture 8

Common Discrete Distributions

p. 4-1 Random Variables

System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models

Statistics for Engineers Lecture 2 Discrete Distributions

Fundamental Tools - Probability Theory II

Topic 9 Examples of Mass Functions and Densities

It can be shown that if X 1 ;X 2 ;:::;X n are independent r.v. s with

Example 1. Assume that X follows the normal distribution N(2, 2 2 ). Estimate the probabilities: (a) P (X 3); (b) P (X 1); (c) P (1 X 3).

Math 151. Rumbos Spring Solutions to Review Problems for Exam 3

Distributions of Functions of Random Variables. 5.1 Functions of One Random Variable

1 Solution to Problem 2.1

HYPERGEOMETRIC and NEGATIVE HYPERGEOMETIC DISTRIBUTIONS

Random variables, Expectation, Mean and Variance. Slides are adapted from STAT414 course at PennState

CS145: Probability & Computing

More Discrete Distribu-ons. Keegan Korthauer Department of Sta-s-cs UW Madison

Discrete Distributions Chapter 6

STAT 3610: Review of Probability Distributions

15 Discrete Distributions

Department of Mathematics

System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models

STAT2201. Analysis of Engineering & Scientific Data. Unit 3

STAT/MATH 395 PROBABILITY II

A Probability Primer. A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes.

Expectations. Definition Let X be a discrete rv with set of possible values D and pmf p(x). The expected value or mean value of X, denoted by E(X ) or

Brief Review of Probability

Department of Mathematics

Common probability distributionsi Math 217 Probability and Statistics Prof. D. Joyce, Fall 2014

Lecture 4: Random Variables and Distributions

Tom Salisbury

Quick Tour of Basic Probability Theory and Linear Algebra

5. Conditional Distributions

Closed book and notes. 120 minutes. Cover page, five pages of exam. No calculators.

Why study probability? Set theory. ECE 6010 Lecture 1 Introduction; Review of Random Variables

Special distributions

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Computer Science, Informatik 4 Communication and Distributed Systems. Simulation. Discrete-Event System Simulation. Dr.

Lecture 6. Probability events. Definition 1. The sample space, S, of a. probability experiment is the collection of all

SDS 321: Introduction to Probability and Statistics

Statistics for Economists Lectures 6 & 7. Asrat Temesgen Stockholm University

Continuous Random Variables

1. Discrete Distributions

Math 407: Probability Theory 5/10/ Final exam (11am - 1pm)

SOLUTION FOR HOMEWORK 12, STAT 4351

Conditional distributions (discrete case)

CDA6530: Performance Models of Computers and Networks. Chapter 2: Review of Practical Random

Order Statistics. The order statistics of a set of random variables X 1, X 2,, X n are the same random variables arranged in increasing order.

CDA5530: Performance Models of Computers and Networks. Chapter 2: Review of Practical Random Variables

ECE353: Probability and Random Processes. Lecture 5 - Cumulative Distribution Function and Expectation

3 Multiple Discrete Random Variables

(Ch 3.4.1, 3.4.2, 4.1, 4.2, 4.3)

Probability and Distributions

Poisson approximations

Lecture 5: Moment generating functions

Discrete Random Variables (cont.) Discrete Distributions the Geometric pmf

Known probability distributions

Lecture 1: August 28

Analysis of Engineering and Scientific Data. Semester

Stat 134 Fall 2011: Notes on generating functions

Transcription:

Mathematical Statistics 1 Math A 6330 Chapter 3 Common Families of Distributions Mohamed I. Riffi Department of Mathematics Islamic University of Gaza September 28, 2015

Outline 1

Subjects of Lecture 04 We will cover the following subjects: Discrete Uniform Distribution Hypergeometric Distribution Binomial Distribution Poisson Distribution Negative Binomial Distribution Geometric Distribution Definition A random variable X is said to have a discrete distribution if the range of X, the sample space, is countable. In most situations, the random variable has integer-valued outcomes.

Subjects of Lecture 04 We will cover the following subjects: Discrete Uniform Distribution Hypergeometric Distribution Binomial Distribution Poisson Distribution Negative Binomial Distribution Geometric Distribution Definition A random variable X is said to have a discrete distribution if the range of X, the sample space, is countable. In most situations, the random variable has integer-valued outcomes.

Discrete Uniform Distribution Definition A random variable X has a discrete uniform(1, N) distribution if P(X = x N) = 1, x = 1, 2,..., N, N where N is a specified integer. This distribution puts equal mass on each of the outcomes 1, 2,..., N.

Remark We use the identities k i = i=1 k(k + 1) 2 and k i 2 = i=1 k(k + 1)(2k + 1) 6 to prove that EX = N + 1 2 EX 2 (N + 1)(2N + 1) = 6 (N + 1)(N 1) Var X = 12

Hypergeometric Distribution Definition Suppose we have a large urn filled with N identical balls M of which are red and N M are green. We select K balls at random. Let X denote the number of red balls in a sample of size K, then X has a hypergeometric distribution given by ( M )( N M ) x K x P(X = x N, M, K) = ( N, x = 0, 1,..., K. K)

Remark Binomial coefficients of the form ( n r) have been defined only if n r, and so the range of X is additionally restricted by the pair of inequalities which can be combined as M x and N M K x, M (N K) x M. Note these identities: ( ) ( ) M M 1 x = M x x 1 and K ( ) ( ) N N 1 = N. K K 1

Remark Binomial coefficients of the form ( n r) have been defined only if n r, and so the range of X is additionally restricted by the pair of inequalities which can be combined as M x and N M K x, M (N K) x M. Note these identities: ( ) ( ) M M 1 x = M x x 1 and K ( ) ( ) N N 1 = N. K K 1

Example Mean and Variance of the Hypergeometric Distribution. EX m = K x m P(X = x) x=0 K ( )( ) /( ) M N M N = x m x K x K x=1 = KM K 1 ( )( ) /( ) M 1 N M N 1 (y + 1) m 1 N y K 1 y K 1 y=0 = KM N E(Y + 1)m 1, where Y hypergeometric(n 1, M 1, K 1).

Example (3.2.1) Suppose that a lot of 25 machine parts is delivered, where a part is considered acceptable only if it passes tolerance. We sample 10 parts and find that none are defective (all are within tolerance). What is the probability of this event if there are 6 defectives in the lot of 25? Solution P(X = 0) = ( 6 19 ) 0)( 10 ) = 0.028. ( 25 10

Example (3.2.1) Suppose that a lot of 25 machine parts is delivered, where a part is considered acceptable only if it passes tolerance. We sample 10 parts and find that none are defective (all are within tolerance). What is the probability of this event if there are 6 defectives in the lot of 25? Solution P(X = 0) = ( 6 19 ) 0)( 10 ) = 0.028. ( 25 10

Binomial Distribution Definition A r.v. X has a Bernoulli(p) distribution, 0 p 1, if { 1 with probability p, X = 0 with probability 1 p. Definition Let Y be the number of successes in n identical Bernoulli trials, each with success probability p, then Y binomial(n, p). The pmf of Y is ( ) n P(Y = y n, p) = p y (1 p) n y, y = 0, 1,..., n. y

Binomial Distribution Definition A r.v. X has a Bernoulli(p) distribution, 0 p 1, if { 1 with probability p, X = 0 with probability 1 p. Definition Let Y be the number of successes in n identical Bernoulli trials, each with success probability p, then Y binomial(n, p). The pmf of Y is ( ) n P(Y = y n, p) = p y (1 p) n y, y = 0, 1,..., n. y

Example n ( ) n EY m = y m p y (1 p) n y y y=0 n ( ) n 1 = np y m 1 p y 1 (1 p) n y y 1 y=1 n 1 ( ) n 1 = np (x + 1) m 1 p x (1 p) n 1 x x x=0 = npe(x + 1) m 1, where X binomial(n 1, p).

Poisson Distribution Definition A random variable X, taking values in the nonnegative integers, has a Poisson(λ) distribution if P(X = x λ) = e λ λ x, x = 0, 1,.... x! Remark P(X = x λ) = e λ x=0 x=0 λ x x! = e λ e λ = 1, where we have used the Taylor series expansion e y = i=0 y i i!.

Poisson Distribution Definition A random variable X, taking values in the nonnegative integers, has a Poisson(λ) distribution if P(X = x λ) = e λ λ x, x = 0, 1,.... x! Remark P(X = x λ) = e λ x=0 x=0 λ x x! = e λ e λ = 1, where we have used the Taylor series expansion e y = i=0 y i i!.

Exercise If X Poisson(λ), then EX = λ EX 2 = λ(λ + 1) Var X = λ M X (t) = e λ(et 1)

Calculating Poisson Probabilities P(X = x) = λ P(X = x 1), x = 1, 2,.... (1) x If Y binomial(n, p), then P(Y = y) = n y + 1 y Set λ = np and, if p is small, we can write p P(Y = y 1). 1 p n y + 1 y p 1 p = np p(y 1) y py λ y. Therefore, P(Y = y) = λ P(Y = y 1), which is Equation (1). y

Calculating Poisson Probabilities To complete the approximation, we should prove that P(X = 0) P(Y = 0). Now, for fixed λ, P(Y = 0) = (1 p) n = (1 λ n )n. lim (1 λ n n )n = e λ = P(X = 0). P(Y = 0) = (1 λ n )n e λ = P(X = 0).

Negative Binomial Definition In a sequence of independent Bernoulli(p) trials, let the random variable X denote the trial at which the rth success occurs. Then ( ) x 1 P(X = x r, p) = p r (1 p) x r, x = r, r + 1,..., r 1 and we say that X has a negative binomial(r, p) distribution. Remark If Y = X r, the number of failures before the rth success, then ( ) r + y 1 P(Y = y) = p y (1 p) y, y = 0, 1,.... y

Negative Binomial Definition In a sequence of independent Bernoulli(p) trials, let the random variable X denote the trial at which the rth success occurs. Then ( ) x 1 P(X = x r, p) = p r (1 p) x r, x = r, r + 1,..., r 1 and we say that X has a negative binomial(r, p) distribution. Remark If Y = X r, the number of failures before the rth success, then ( ) r + y 1 P(Y = y) = p y (1 p) y, y = 0, 1,.... y

Remark Note that ( ) r + y 1 y ( ) r = ( 1) y y y ( r)( r 1) ( r y + 1) = ( 1). (y)(y 1) (2)(1) Therefore, we may write the pmf of Y as P(Y = y) = ( 1) y ( r y ) p r (1 p) y, y = 0, 1,.... Note that x ( ) ( ) x 1 x = r. r 1 r

Example (Moments of Negative Binomial) Compute the expected value and the variance of a negative binomial random variable with parameters r and p. ( ) x 1 EX k = x k p r (1 p) x r r 1 x=r = r ( ) x x k 1 p r+1 (1 p) x r p r x=r = r ( ) y 1 (y 1) k 1 p r+1 (1 p) y (r+1) p r y=r+1 = r p E(Y 1)k 1, where Y negative binomial(r + 1, p).

Example (Moments of Negative Binomial) Compute the expected value and the variance of a negative binomial random variable with parameters r and p. ( ) x 1 EX k = x k p r (1 p) x r r 1 x=r = r ( ) x x k 1 p r+1 (1 p) x r p r x=r = r ( ) y 1 (y 1) k 1 p r+1 (1 p) y (r+1) p r y=r+1 = r p E(Y 1)k 1, where Y negative binomial(r + 1, p).

Remark Since Y = X r, it follows that r(1 p) EY = EX r = p r(1 p) Var Y = Var X = p 2 If we define µ = r(1 p) p, then EY = µ and Var Y = µ + 1 r µ2. If r and p 1 such that r(1 p) λ, 0 < λ <, then EY = Var Y = r(1 p) λ, p r(1 p) p 2 λ.

Remark Since Y = X r, it follows that r(1 p) EY = EX r = p r(1 p) Var Y = Var X = p 2 If we define µ = r(1 p) p, then EY = µ and Var Y = µ + 1 r µ2. If r and p 1 such that r(1 p) λ, 0 < λ <, then EY = Var Y = r(1 p) λ, p r(1 p) p 2 λ.

Example Find the expected value and the variance of the number of times one must throw a die until the outcome 1 has occurred 4 times. Solution Since the random variable of interest is a negative binomial with parameters r = 4 and p = 1 6, it follows that EX = 24 Var X = 4 ( ) 5 6 ( 1 2 = 120. 6)

Example Find the expected value and the variance of the number of times one must throw a die until the outcome 1 has occurred 4 times. Solution Since the random variable of interest is a negative binomial with parameters r = 4 and p = 1 6, it follows that EX = 24 Var X = 4 ( ) 5 6 ( 1 2 = 120. 6)

Inverse Binomial Sampling Example (3.2.6) suppose that in a population of fruit flies we are interested in the proportion having vestigial wings and decide to sample until we have found 100 such flies. The probability that we will have to examine at least N flies is ( ) x 1 P(X N) = p 100 (1 p) x 100 99 x=n = 1 N 1 x=100 ( x 1 99 ) p 100 (1 p) x 100. For given p and N, we can evaluate this expression to determine how many fruit flies we are likely to look at.

Inverse Binomial Sampling Example (3.2.6) suppose that in a population of fruit flies we are interested in the proportion having vestigial wings and decide to sample until we have found 100 such flies. The probability that we will have to examine at least N flies is ( ) x 1 P(X N) = p 100 (1 p) x 100 99 x=n = 1 N 1 x=100 ( x 1 99 ) p 100 (1 p) x 100. For given p and N, we can evaluate this expression to determine how many fruit flies we are likely to look at.

Geometric Distribution Definition The geometric distribution is the simplest of the waiting time distributions and is a special case of the negative binomial distribution, where r = 1. If X geometric(p), then its pdf is P(X = x p) = p(1 p) x 1, x = 1, 2,.... Remark The mean and variance of X geometric(p) are respectively EX = 1 p and Var X = 1 p p 2.

Geometric Distribution Definition The geometric distribution is the simplest of the waiting time distributions and is a special case of the negative binomial distribution, where r = 1. If X geometric(p), then its pdf is P(X = x p) = p(1 p) x 1, x = 1, 2,.... Remark The mean and variance of X geometric(p) are respectively EX = 1 p and Var X = 1 p p 2.

Memoryless Property of Geometric Distribution Remark X geometric(p) has the memoryless property; that is, P(X > s X > t) = P(X > s t), s > t > 0. (2) To establish (2), note for any integer n that P(X > n) = P(no success in n trials) = (1 p) n. P(X > s X > t) = P(X > s, X > t) P(X > t) = P(X > s) P(X > t) = (1 p) s t = P(X > s t).

Memoryless Property of Geometric Distribution Remark X geometric(p) has the memoryless property; that is, P(X > s X > t) = P(X > s t), s > t > 0. (2) To establish (2), note for any integer n that P(X > n) = P(no success in n trials) = (1 p) n. P(X > s X > t) = P(X > s, X > t) P(X > t) = P(X > s) P(X > t) = (1 p) s t = P(X > s t).

Example (3.2.7) The geometric distribution is sometimes used to model lifetimes or time until failure of components. For example, if the probability is.001 that a light bulb will fail on any given day, then the probability that it will last at least 30 days is P(X > 30) =.001(1.001) x 1 = (.999) 30 =.970. x=30