Discrete Distributions

Similar documents
Statistics for Economists. Lectures 3 & 4

Relationship between probability set function and random variable - 2 -

Chapter 2: Discrete Distributions. 2.1 Random Variables of the Discrete Type

STAT/MATH 395 A - PROBABILITY II UW Winter Quarter Moment functions. x r p X (x) (1) E[X r ] = x r f X (x) dx (2) (x E[X]) r p X (x) (3)

Mathematical Statistics 1 Math A 6330

Continuous Distributions

Things to remember when learning probability distributions:

Introduction to Statistics. By: Ewa Paszek

STAT/MATH 395 PROBABILITY II

Random Variables Example:

Discrete random variables and probability distributions

15 Discrete Distributions

Probability Distributions Columns (a) through (d)

Binomial and Poisson Probability Distributions

Chapter 3. Discrete Random Variables and Their Probability Distributions

MAT X (Spring 2012) Random Variables - Part I

Discrete Distributions Chapter 6

4 Moment generating functions

1 Review of Probability

STAT 3610: Review of Probability Distributions

System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models

Topic 3: The Expectation of a Random Variable

Random Variables. Definition: A random variable (r.v.) X on the probability space (Ω, F, P) is a mapping

Week 2. Review of Probability, Random Variables and Univariate Distributions

Chapter 3. Discrete Random Variables and Their Probability Distributions

Chapter 5. Chapter 5 sections

Distributions of Functions of Random Variables. 5.1 Functions of One Random Variable

Northwestern University Department of Electrical Engineering and Computer Science

A Probability Primer. A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes.

1.6 Families of Distributions

Introduction to Probability and Statistics Slides 3 Chapter 3

Discrete Random Variables

Brief Review of Probability

Chapter 2. Discrete Distributions

Lecture 3. Discrete Random Variables

Suppose that you have three coins. Coin A is fair, coin B shows heads with probability 0.6 and coin C shows heads with probability 0.8.

Expectation, variance and moments

Discrete Distributions

Topic 3: The Expectation of a Random Variable

Chapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University

Contents 1. Contents

Stochastic Processes - lesson 2

Random variables, Expectation, Mean and Variance. Slides are adapted from STAT414 course at PennState

Lecture 5: Moment generating functions

Chapter 4 : Discrete Random Variables

EE/CpE 345. Modeling and Simulation. Fall Class 5 September 30, 2002

1.1 Review of Probability Theory

3 Modeling Process Quality

Random variables. DS GA 1002 Probability and Statistics for Data Science.

EXAM. Exam #1. Math 3342 Summer II, July 21, 2000 ANSWERS

Polytechnic Institute of NYU MA 2212 MIDTERM Feb 12, 2009

Topic 3 - Discrete distributions

Copyright c 2006 Jason Underdown Some rights reserved. choose notation. n distinct items divided into r distinct groups.

Slides 8: Statistical Models in Simulation

System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models

Fundamental Tools - Probability Theory II

Moments. Raw moment: February 25, 2014 Normalized / Standardized moment:

Chapter 8. Some Approximations to Probability Distributions: Limit Theorems

Lecture 6. Probability events. Definition 1. The sample space, S, of a. probability experiment is the collection of all

Continuous RVs. 1. Suppose a random variable X has the following probability density function: π, zero otherwise. f ( x ) = sin x, 0 < x < 2

, find P(X = 2 or 3) et) 5. )px (1 p) n x x = 0, 1, 2,..., n. 0 elsewhere = 40

Discrete Random Variables

Review 1: STAT Mark Carpenter, Ph.D. Professor of Statistics Department of Mathematics and Statistics. August 25, 2015

Lecture 6: Special probability distributions. Summarizing probability distributions. Let X be a random variable with probability distribution

FINAL EXAM: Monday 8-10am

Bernoulli Trials, Binomial and Cumulative Distributions

Moment Generating Functions

SUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416)

Probability and Statistics Concepts

Common probability distributionsi Math 217 Probability and Statistics Prof. D. Joyce, Fall 2014

Continuous Probability Spaces

Exam 3, Math Fall 2016 October 19, 2016

Special Discrete RV s. Then X = the number of successes is a binomial RV. X ~ Bin(n,p).

An-Najah National University Faculty of Engineering Industrial Engineering Department. Course : Quantitative Methods (65211)

STAT2201. Analysis of Engineering & Scientific Data. Unit 3

ELEG 3143 Probability & Stochastic Process Ch. 2 Discrete Random Variables

PROBABILITY DISTRIBUTION

Uses of Asymptotic Distributions: In order to get distribution theory, we need to norm the random variable; we usually look at n 1=2 ( X n ).

MATH Notebook 5 Fall 2018/2019

Discrete Random Variables

Probability and Distributions

1 Random Variable: Topics

Lecture 13. Poisson Distribution. Text: A Course in Probability by Weiss 5.5. STAT 225 Introduction to Probability Models February 16, 2014

Expectations. Definition Let X be a discrete rv with set of possible values D and pmf p(x). The expected value or mean value of X, denoted by E(X ) or

p. 4-1 Random Variables

1 Solution to Problem 2.1

Introduction to Statistical Data Analysis Lecture 3: Probability Distributions

SOLUTION FOR HOMEWORK 12, STAT 4351

Chapter 1 Statistical Reasoning Why statistics? Section 1.1 Basics of Probability Theory

Chapter 2. Random Variable. Define single random variables in terms of their PDF and CDF, and calculate moments such as the mean and variance.

Discrete Probability Distributions

Quick review on Discrete Random Variables

Test Problems for Probability Theory ,

Actuarial Science Exam 1/P

Conditional Probability

Probability, Random Processes and Inference

To find the median, find the 40 th quartile and the 70 th quartile (which are easily found at y=1 and y=2, respectively). Then we interpolate:

Probability- the good parts version. I. Random variables and their distributions; continuous random variables.

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Math Bootcamp 2012 Miscellaneous

Transcription:

Chapter 2 Discrete Distributions 2.1 Random Variables of the Discrete Type An outcome space S is difficult to study if the elements of S are not numbers. However, we can associate each element/outcome of S with a number. Def. Given a random experiment with an outcome space S, a function X that assigns one and only one real number X(s) = x to each element s in S is called a random variable. The space of X is the set of real numbers {x X(s) = x, s S}. When the elements of S are real numbers, the identical function X(s) = s is itself a random variable with the space S. Ex (Ex 2, p52, scanned file) With random variables, the probabilities of outcomes/events in S may be studied by the probabilities of random variables (e.g. P (X = x), P (a < X b), etc.) If the outcome space S has finitely many elements or countable many elements, i.e. the elements of S can be listed in ordered: S = {s 1, s 2, }, then S is called a discrete outcome space, and a random variable X is called a random variable of the discrete type. For a random variable X of discrete type, the probability P (X = x) is frequently denoted by f(x), and is called the probability mass function (p.m.f.). 15

16 CHAPTER 2. DISCRETE DISTRIBUTIONS Def. A probability mass function (p.m.f.) f(x) of a discrete random variable X is a function that satisfies the following properties: (Let S be the space of r.v. X.) 1. f(x) > 0, x S; 2. x S f(x) = 1 3. P (X A) = x A f(x), where A S. We denote f(x) = 0 when x / S. Ex Ex 3, p.54 (scanned file) The p.m.f. may be described by bar graph or probability histogram (Fig 2.1-1, p.54, scanned file). Hypergeometric distribution: Consider a collection of N = N 1 + N 2 objects; N 1 of them belong to class 1; N 2 of them belong to class 2. A collection of n objects is selected from these N objects without replacement. The probability that exactly x objects of the n objects are in class 1, and n x objects are in class 2, is ) f(x) = P (X = x) = ( N1 )( N2 x n x ( N x ). Ex Ex 4-Ex 5, p55-p56 (scanned file) Homework 2.1 1, 7, 9, 11, 13 Attachment: Scanned textbook pages of Section 2-1

2.2. MATHEMATICAL EXPECTATION 17 2.2 Mathematical Expectation Ex 1 (p.61, scanned file) Def. Let f(x) be the p.m.f. of the r.v. X of the discrete type with space S (i.e. f : S R). Let u(x) be a function with space S (i.e. u : S R). The sum E[u(X)] := x S u(x)f(x) is called the mathematical expectation or the expected value of the function u(x). The definition of E[u(X)] requires that the sum converges absolutely, that is, u(x) f(x) converges to a finite number. x S Ex Let f(x) be the p.m.f. of the r.v. X. Then E(X) = x S xf(x), E(X n ) = x S x n f(x). Ex 2 (p.63, scanned file) Thm 2.1. The mathematical expectation E satisfies the following properties: 1. E(c) = c for constant c; 2. E[cu(X)] = ce[u(x)] for constant c and function u(x). 3. E[c 1 u 1 (X)+c 2 u 2 (X)] = c 1 E[u 1 (X)]+c 2 E[u 2 (X)] for constants c 1 and c 2, and functions u 1 (X) and u 2 (X). Ex (Exercise 2.2-2, p.66) Let the r.v. X have the p.m.f. f(x) = ( x + 1)2, x = 1, 0, 1. 9 Compute E(X), E(X 2 ), and E(3X 2 2X + 4). Homework 2.2 1, 3, 7 Attachment: Scanned textbook pages of Section 2-2

18 CHAPTER 2. DISCRETE DISTRIBUTIONS 2.3 The Mean, Variance, and Standard Deviation Let X be a random variable with space S = {u 1, u 2,, u k }. Let f : S [0, 1] be the p.m.f. of X. The mean of X (or of its distribution) is µ := E[X] = x S xf(x) = u 1 f(u 1 ) + u 2 f(u 2 ) + + u k f(u k ). The variance of X (or of its distribution) is Var(X) := E[(X µ) 2 ] = x S(x µ) 2 f(x) = (u 1 µ) 2 f(u 1 ) + (u 2 µ) 2 f(u 2 ) + + (u k µ) 2 f(u k ). The standard deviation of X is σ = Var(X) = E[(X µ) 2 ]. One can show that σ 2 = Var(X) = E[(X µ) 2 ] = E[X 2 2µX + µ 2 ] = E[X 2 ] 2µ 2 + µ 2 = E[X 2 ] µ 2 = x 2 f(x) µ 2. x S Ex 1, p69 (scanned file.) The variance and standard deviation measure how dispersed the data are from the mean. Ex 3, p70 (scanned file.) Ex 4, p70, the mean and the variance of a uniform distribution. The rth moment of the distribution X about the origin is E[X r ] = x S x r f(x).

2.3. THE MEAN, VARIANCE, AND STANDARD DEVIATION 19 The rth moment of the distribution X about b is E[(X b) r ] = x S(x b) r f(x). The rth factorial moment of the distribution X is E[(X) r ] = E[X(X 1)(X 2) (X r + 1)]. If we perform a random experiment n times and obtain n observed values of the random variable say, x 1, x 2,, x n. This collection is called a sample. We actually create a probability distribution with weight 1/n on each of these x values, called the empirical distribution. The sample mean of the empirical distribution is x := 1 n n x i. i=1 It gives an estimation of µ = E[X]. The variance of the empirical distribution is v = E[(X x) 2 ] = 1 n n (x i x) 2. i=1 HOWEVER, Var(X) = σ 2 is better estimated by the sample variance s 2 := ( ) n v = 1 n 1 n 1 Here s is the sample standard deviation. Ex (Exercise 2.3-10, p76, see scanned file.) n (x i x) 2. i=1 Homework 2.3 1, 3, 5, 9, 15 Attachment: Scanned textbook pages of Section 2-3

20 CHAPTER 2. DISCRETE DISTRIBUTIONS 2.4 Bernoulli Trials and the Binomial Distribution A Bernoulli experiment is an experiment with only two outcomes say, success or failure (e.g. female or male, nondefective or defective). Let p to denote the probability of success, and q = 1 p the probability of failure. Let the r.v. X denote as follow: The p.m.f. of X is X(success) = 1, X(failure) = 0. f(1) = p, f(0) = q = 1 p. We say that X has a Bernoulli distribution. The expected value of X is The variance of X is µ = E(X) = 1(p) + 0(1 p) = p. σ 2 = Var(X) = E[(X µ) 2 ] The standard deviation of X is = (1 p) 2 (p) + (0 p) 2 (1 p) = p(1 p). σ = p(1 p). Ex 1, p78 (file). A sequence of Bernoulli trials occurs when 1. A Bernoulli (success-failure) experiment is performed n times. 2. The trials are independent. 3. The probability of success on each trial is a constant p; the probability of failure is q = 1 p. Ex 4, p79 (file).

2.4. BERNOULLI TRIALS AND THE BINOMIAL DISTRIBUTION 21 In a sequence of n Bernoulli trials, let the random variable X denote the number of successes in n trials. The space of X is {0, 1,, n}. If x successes occur, then we have x successes in n trials and the remaining n x trials are failures. The p.m.f. of X is f(x) = P (X = x) = ( ) n p x (1 p) n x. x These probabilities are called binomial probabilities, and the r.v. X is said to have a binomial distribution. A binomial distribution will be denoted by b(n, p). The numbers n and p are called the parameters of the binomial distribution; they correspond to the number n of independent trials and the probability p of success on each trial. Ex 6, Fig 2.4-1, p80-81 (file) Ex 7, p80 (file). For a random variable X, the cumulative distribution function (or the distribution function) is defined as: F (x) := P (X x), < X <. Ex (Fig 2.4-2 on p82, file) The distribution function for the b(10, 0.8) distribution. Ex 9, p82 (file). Let X be the r.v. with the binomial distribution b(n, p). The sum of p.m.f. over the space of X is 1: n x=0 ( ) n p x (1 p) n x = [p + (1 p)] n = 1 x It can be applied to compute the mean of X. (true for all n).

22 CHAPTER 2. DISCRETE DISTRIBUTIONS Prop 2.2 (Binomial distribution X b(n, p)). p.m.f.: Mean µ = E(X): Variance σ 2 = E(X 2 ) µ 2 : f(x) = ( ) n x p x (1 p) n x µ = np σ 2 = np(1 p). Standard deviation: σ = np(1 p) Distribution function: Table II in the appendix Ex 10 and 11, p84-85 (file). Homework 2.4 1, 3, 5, 11, 15, 21 Attachment: Scanned textbook pages of Section 2-4

2.5. THE MOMENT-GENERATING FUNCTION 23 2.5 The Moment-Generating Function The mean, variance, and standard deviation are important characteristics of a distribution. However, it may be difficult to compute E(X) or E(X 2 ) for some distributions (e.g. binomial distribution.) We define a new function of t that will help use generate the moments of a distribution. Def. Let X be a r.v. of discrete type with p.m.f. f(x) and space S. The function of t: M(t) := E[e tx ] = x S e tx f(x), h < t < h, is called the moment-generating function of X (abbreviated as m.g.f.), provided that M(t) converges for h < t < h. Prop 2.3. Properties of m.g.f. M(t) of X: 1. If the space S = {b 1, b 2, b 3, }, then M(t) = e tb 1 f(b 1 ) + e tb 2 f(b 2 ) + e tb 3 f(b 3 ) + Thus, the coefficient of e tb i is the probability f(b i ) = P (X = b i ). 2. Two random variables have the same m.g.f. if and only if they have the same p.m.f. 3. The derivatives of M(t) at 0 generate all E[X r ]: M(t) = E[e tx ] = x S e tx f(x), M(0) = E[1] = 1; M (t) = E[Xe tx ] = x S xe tx f(x), M (0) = E[X]; M (t) = E[X 2 e tx ] = x S x 2 e tx f(x), M (0) = E[X 2 ]; M (r) (t) = E[X r e tx ] = x S x r e tx f(x), M (r) (0) = E[X r ].. Ex 1, p90: Suppose X has the m.g.f. M(t) = e t ( 3 6 ) + e2t ( 2 6 ) + e3t ( 1 6 ). Describe the space and the p.m.f. of X.

24 CHAPTER 2. DISCRETE DISTRIBUTIONS Ex 2, p90: Describe X if the r.v. X has the m.g.f. M(t) = et /2 1 e t, t < ln 2. /2 Ex 3, p91: Use the m.g.f. of X to find the mean, the variance, and the standard deviation of the binomial distribution X b(n, p) with p.m.f.: ( ) n f(x) = p x (1 p) n x, x = 0, 1, 2,, n, x (Answer: M(t) = [(1 p) + pe t ] n. See scanned file.)....................................................................... Let r be a fixed positive integer. Let the r.v. X denote the number of Bernoulli trials needed to observe the r-th successes. Then P (X = x) is the probability to see the r-th success in the x-th trial, that is, (r 1) successes in the first (x 1) trials and success in the last trial. The p.m.f. of X is ( ) x 1 g(x) = P (X = x) = p r 1 (1 p) (x 1) (r 1) p = r 1 ( x 1 r 1 ) p r (1 p) x r, x = r, r + 1,. We say that X has a negative binomial distribution (parameters r, p). Binomial distribution: Negative binomial distribution: Fix the number of Bernoulli trials. Observe the number of successes. Fix the number of successes. Observe the number of Bernoulli trials needed. Prop 2.4 (Negative binomial distribution X). p.m.f.: g(x) = ( x 1 r 1) p r (1 p) x r, x = r, r + 1, m.g.f.: M(t) = (pe t ) r [1 (1 p)e t ] r where (1 p)e t < 1. Mean: Variance: µ = E(X) = r p σ 2 = r(1 p) p 2.

2.5. THE MOMENT-GENERATING FUNCTION 25 The negative binomial distribution for r = 1 is called a geometric distribution, since the p.m.f. of X consists of terms of a geometric series, namely g(x) = p(1 p) x 1, x = 1, 2, 3,........................................................................ Ex 4, p93 (scanned file) Ex (HW 16, p98) (scanned file) Ex 5, p94 (scanned file) Ex 6, p95 (scanned file) Homework 2.5 1, 7, 9, 15 Attachment: Scanned textbook pages of Section 2-5

26 CHAPTER 2. DISCRETE DISTRIBUTIONS 2.6 The Poisson Distribution Def. In counting the number of changes that occur in a time interval, we have an approximate Poisson process with parameter λ > 0 if: 1. The number of changes occurring in nonoverlapping intervals are independent. 2. The probability of exactly one change occurring in a sufficiently short interval of length h is approximately λh. 3. The probability of two or more changes occurring in a sufficiently short interval is essentially zero. Let the r.v. X denote the number of changes that occur in an interval of length 1. To get P (X = x), we divide the unit interval into n subintervals with equal length 1/n. We have a sequence of n Bernoulli trials with probability p λ/n. Therefore, P (X = x) Take limit for n : P (X = x) = lim n n! x!(n x)! n! x!(n x)! ( ) λ x ( 1 λ n x. n n) ( ) λ x ( 1 λ ) n x = λx e λ. n n x! Prop 2.5 (Poisson distribution X with parameter λ > 0). p.m.f.: f(x) = λx e λ x!, x = 0, 1, 2, m.g.f.: M(t) = E(e tx ) = e λ(et 1) Mean: Variance: Distribution function: µ = E(X) = λ σ 2 = E(X 2 ) µ 2 = λ Table III in the appendix Ex 1, 2, 3, p101-103, Table III in the appendix (scanned file)

2.6. THE POISSON DISTRIBUTION 27 If events in a Poisson process occur at a mean rate of λ per unit, then the number of occurrences say, X in the interval of length t has the Poisson p.m.f. f(x) = (λt)x e λt, x = 0, 1, 2, x! It just likes that we treat the interval of length t as if it were the unit interval with mean λt instead of λ. Ex 4, 5, p103-104, Table III in the appendix (scanned file) When n is large and p is small, the binomial distribution b(n, p) may be approximated by the Poisson distribution with λ = np. Ex Fig 2.6-3, p105 (scanned file) Homework 2.6 1, 7, 9, 15 Attachment: Scanned textbook pages of Section 2-6