Chapter 3. Discrete Random Variables and Their Probability Distributions

Similar documents
Chapter 3. Discrete Random Variables and Their Probability Distributions

Mathematical Statistics 1 Math A 6330

Discrete Distributions

Lecture 6. Probability events. Definition 1. The sample space, S, of a. probability experiment is the collection of all

Binomial and Poisson Probability Distributions

ELEG 3143 Probability & Stochastic Process Ch. 2 Discrete Random Variables

Random Variables Example:

Statistics for Economists. Lectures 3 & 4

Topic 3 - Discrete distributions

3.4. The Binomial Probability Distribution

Introduction to Statistical Data Analysis Lecture 3: Probability Distributions

Question Points Score Total: 137

Things to remember when learning probability distributions:

STAT2201. Analysis of Engineering & Scientific Data. Unit 3

Relationship between probability set function and random variable - 2 -

1 Bernoulli Distribution: Single Coin Flip

Known probability distributions

Week 6, 9/24/12-9/28/12, Notes: Bernoulli, Binomial, Hypergeometric, and Poisson Random Variables

Chapter 2: Discrete Distributions. 2.1 Random Variables of the Discrete Type

Math/Stat 352 Lecture 8

Suppose that you have three coins. Coin A is fair, coin B shows heads with probability 0.6 and coin C shows heads with probability 0.8.

Notes for Math 324, Part 17

STAT509: Discrete Random Variable

Statistics for Engineers Lecture 2 Discrete Distributions

3 Modeling Process Quality

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH A Test #2 June 11, Solutions

Closed book and notes. 120 minutes. Cover page, five pages of exam. No calculators.

Bernoulli and Binomial Distributions. Notes. Bernoulli Trials. Bernoulli/Binomial Random Variables Bernoulli and Binomial Distributions.

Discrete Random Variable Practice

Discrete Distributions

Expectations. Definition Let X be a discrete rv with set of possible values D and pmf p(x). The expected value or mean value of X, denoted by E(X ) or

Chapter 3 Discrete Random Variables

Discrete probability distributions

Lecture 14. Text: A Course in Probability by Weiss 5.6. STAT 225 Introduction to Probability Models February 23, Whitney Huang Purdue University

Lecture 3. Discrete Random Variables

System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models

Random Variables. Definition: A random variable (r.v.) X on the probability space (Ω, F, P) is a mapping

15 Discrete Distributions

Stat Lecture 20. Last class we introduced the covariance and correlation between two jointly distributed random variables.

p. 4-1 Random Variables

Computer Science, Informatik 4 Communication and Distributed Systems. Simulation. Discrete-Event System Simulation. Dr.

Special Discrete RV s. Then X = the number of successes is a binomial RV. X ~ Bin(n,p).

Lecture 08: Poisson and More. Lisa Yan July 13, 2018

B.N.Bandodkar College of Science, Thane. Subject : Computer Simulation and Modeling.

STAT 509 Section 3.4: Continuous Distributions. Probability distributions are used a bit differently for continuous r.v. s than for discrete r.v. s.

Discrete random variables and probability distributions

Binomial random variable

Chapter 3: Discrete Random Variable

Discrete Probability Distributions

Common Discrete Distributions

Geometric Distribution The characteristics of a geometric experiment are: 1. There are one or more Bernoulli trials with all failures except the last

Contents 1. Contents

II. The Binomial Distribution

System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models

An-Najah National University Faculty of Engineering Industrial Engineering Department. Course : Quantitative Methods (65211)

STA 4321/5325 Solution to Extra Homework 1 February 8, 2017

Chapter 3: Discrete Random Variables

Discrete Distributions

Topic 3: The Expectation of a Random Variable

Bayes Theorem and Hypergeometric Distribution

(y 1, y 2 ) = 12 y3 1e y 1 y 2 /2, y 1 > 0, y 2 > 0 0, otherwise.

Chapter 5 Class Notes

Chapter 4a Probability Models

Introduction to Probability and Statistics Slides 3 Chapter 3

A Probability Primer. A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes.

BINOMIAL DISTRIBUTION

CS 237: Probability in Computing

37.3. The Poisson Distribution. Introduction. Prerequisites. Learning Outcomes

Chapter 3 Common Families of Distributions

Slides 8: Statistical Models in Simulation

(b). What is an expression for the exact value of P(X = 4)? 2. (a). Suppose that the moment generating function for X is M (t) = 2et +1 3

Lecture 13. Poisson Distribution. Text: A Course in Probability by Weiss 5.5. STAT 225 Introduction to Probability Models February 16, 2014

(Ch 3.4.1, 3.4.2, 4.1, 4.2, 4.3)

Continuous-Valued Probability Review

GEOMETRIC -discrete A discrete random variable R counts number of times needed before an event occurs

Discrete Distributions Chapter 6

(Ch 3.4.1, 3.4.2, 4.1, 4.2, 4.3)

Stat 515 Midterm Examination II April 4, 2016 (7:00 p.m. - 9:00 p.m.)

Review for the previous lecture

Discrete & Continuous Probability Distributions Sunu Wibirama

STAT 345 Spring 2018 Homework 4 - Discrete Probability Distributions

Probability and Statistics for Engineers

Bernoulli Trials, Binomial and Cumulative Distributions

POISSON PROCESSES 1. THE LAW OF SMALL NUMBERS

Notes on Continuous Random Variables

Discrete Probability Distributions

1. Discrete Distributions

Closed book and notes. 60 minutes. Cover page and four pages of exam. No calculators.

Random Variable And Probability Distribution. Is defined as a real valued function defined on the sample space S. We denote it as X, Y, Z,

Chapter 5. Chapter 5 sections

Statistics 427: Sample Final Exam

Stochastic Models of Manufacturing Systems

Each trial has only two possible outcomes success and failure. The possible outcomes are exactly the same for each trial.

Probability Distributions Columns (a) through (d)

Topic 3: The Expectation of a Random Variable

Math st Homework. First part of Chapter 2. Due Friday, September 17, 1999.

b. ( ) ( ) ( ) ( ) ( ) 5. Independence: Two events (A & B) are independent if one of the conditions listed below is satisfied; ( ) ( ) ( )

14.2 THREE IMPORTANT DISCRETE PROBABILITY MODELS

Analysis of Engineering and Scientific Data. Semester

Probability and Statistics

Transcription:

Chapter 3. Discrete Random Variables and Their Probability Distributions 1

3.4-3 The Binomial random variable The Binomial random variable is related to binomial experiments (Def 3.6) 1. The experiment consists of n identical and independent trials. 2. Each trial results in one of two outcomes (concerned with the r.v. Y of interest). We call one outcome a success S and the other a failure F. Here, success is merely a name for one of the two possible outcomes on a single trial of an experiment. 3. The probability of success on a single trial is equal to p and remains the same from trial to trial. The probability of a failure is equal to q(= 1 p). 4. The random variable of interest is Y, the number of successes observed during the n trials. (Example 3.5) Reading 2

Before giving the definition of the binomial p.m.f., try to derive it by using the samplepoint approach. How? 1. Each sample point in the sample space can be characterized by an n-tuple involving the letters S and F, corresponding to success and failure: Ex) SSF SF SF F F F SS... SF }{{}. n positions(n trials) The letter in the i-th position(proceeding from left to right) indicates the outcome of the i-th trial. 2. Consider a particular sample point corresponding to y successes and contained in the numerical event Y = y. Ex) SSSS }{{... SSS} F F F... F F y n-y }{{}. This sample point represents the intersection of n independent events in which there were y successes followed by (n y) failures. 3

3. If the probability of success(s) is p, this probability is unchanged from trial to trial because the trials are independent. So, the probability of the sample point in 2 is Ex) pppp... ppp }{{} y terms qqq... qq }{{} = py q n y. n-y terms Every other sample point in the event Y = y can be represented as an n-tuple containing y S s and (n-y) F s in some order. probability p y q n y.. Any such sample point also has 4. The number of distinct n-tuples that contain y S s and (n ( y) F s is n) n! = y y!(n y)!. 5. The event (Y = y) is made up of ( ) n sample points, each with probability p y q n y, and that the binomial probability distribution is p(y) = ( n) p y q n y. y y 4

(Def 3.7) A random variable Y is said to have a binomial distribution with the parameters n trials and success probability p (in the binomial experiment) (i.e., Y b(n, p)) if and only if p(y) = ( n) p y q n y, y where q = 1 p, y = 0, 1, 2,..., n and 0 p 1. How about Y b(1, p)? (Question) Does p(y) in (Def 3.7) satisfy the necessary properties in (Theorem 3.1)? (Example 3.7) Suppose that a lot of 5000 electrical fuses contains 5% defectives. If a sample of five fuses is tested, find the probability of observing at least one defective. 5

(Exercise 3.39) A complex electronic system is built with a certain number of backup components in its subsystems. One subsystem has four identical components, each with a probability of.2 of failing in less than 1000 hours. The system will operate if any two of the four components are operating. Assume that the components operate independently. (a) Find the probability that exactly two of the four components last longer than 1000 hours. (b) Find the probability that the subsystem operates longer than 1000 hours. (Theorem 3.7) Let Y be a binomial random variable based on n trials and success probability p. Then, (Proof) µ = E(Y ) = np and σ 2 = V (Y ) = npq (Example 3.7) Mean and Variance (Exercise 3.39) Mean and Variance 6

3.5 The Geometric random variable The geometric random variable is related to the following experiments 1. The experiment consists of identical and independent trials, but the number of trials is not fixed. 2. Each trial results in one of two outcomes (concerning with the r.v, Y ), a success and a failure. 3. The probability of success on a single trial is equal to p and remains the same from trial to trial. The probability of a failure is equal to q(= 1 p). 4. However, the random variable of interest Y is the number of the trial on which the first success occurs, not the number of successes that occur in n trials. So, the experiment could end with the first trial if a success is observed on the very first trial, or the experiment could go on indefinitely!!. 7

(Def 3.8) A random variable Y is said to have a geometric probability distribution with the parameter p, success probability (i.e., Y Geo(p)) if and only if p(y) = q y 1 p, where q = 1 p, y = 1, 2, 3..., and 0 p 1. (Question) Does p(y) in (Def 3.8) satisfy the necessary properties in (Theorem 3.1)? (Exercise 3.67) Suppose that 30% of the applicants for a certain industrial job possess advanced training in computer programming. Applicants are interviewed sequentially and are selected at random from the pool. Find the probability that the first applicant with advanced training in programming is found on the fifth interview. (Example) A basket player can make a free throw 60% of the time. Let X be the minimum number of free throws that this player must attempt to make first shot. What is P (X = 5)? 8

(Theorem 3.8) Let Y be a binomial random variable with a geometric distribution, µ = E(Y ) = 1 p and σ2 = V (Y ) = 1 p p 2 (Exercise 3.67 and Example above) Mean and Variance [Memoryless property] CDF of Y Geo(p) : F Y (a) = P (Y a) = P (Y > a + b Y > a) = P (Y > b) : given that the first success has not yet occurred, the probability of the number of additional trials does not depend on how many failures has been observed. 9

3.7 The Hypergeometric random variable The hypergeometric random variable is related to the following experiments 1. In the population of N elements there are elements of two distinct types (success and failure) 3. A sample of size n is randomly selected without replacement from a population of N elements 4. The random variable of interest is Y, the number of success in the sample 2. Among N elements r elements can be classified as success and N r elements can be classified as failture. (Example) A bowl contains N chips, of which N 1 are white, and N 2 are green chips. Randomly select n chips from the bowl without replacement. Let Y be the number of white chips chosen. What is P (Y = y)? 10

(Def 3.10) A random variable Y is said to have a hypergeometric probability distribution with the parameters N andr if and only if p(y) = ( )( ) r N r y n y ( ) N, n where y is an integer 0, 1, 2,..., n, subject to the restrictions y r and n y N r. (Question) Does p(y) in (Def 3.10) satisfy the necessary properties in (Theorem 3.1)? (Hint) Use the following facts : ( a b) = 0 if b > a, n ( r )( N r ) = i n i i=0 ( N n ). (Theorem 3.10) Let Y be a random variable with a hypergeometric distribution, µ = E(Y ) = nr N and ( ) ( ) ( ) σ 2 r N r N n = V (Y ) = n. N N N 1 11

(Exercise 3.103) A warehouse contains ten printing machines, four of which are defective. A company selects five of the machines at random, thinking all are in working condition. What is the probability that all five of the machines are nondefective? [Relationship between Binomial distribution and Hypergeometric distribution] When N is large, n is relatively small and r/n is held constant and equal to p, the following holds: lim N where r/n = p. ( )( ) r N r y n y ( N n ) = ( n) p y (1 p) n y y 12

We learned the following discrete random variable and their probability distributions (p.m.f.): 1) Discrete uniform probability distribution 2) Bernoulli probability distribution 3) Binomial probability distribution 4) Geometric probability distribution 5) Hypergeometric probability distribution The experiments in 2)-5) has two outcomes concerned with the r.v. Y, for example success and failure. Now we will learn how to model counting data (number of times a particular event occurs) : Poisson r.v. and its probability distribution. 13

3.8 The Poisson random variable The Poisson r.v. often provides a good model for the probability distribution of the number Y of (rare) events that occur in a fixed space, time interval, volume, or any other dimensions. (Example) the number of automobile accidents, or other types of accidents in a given unit of time. (Example) the number of prairie dogs found in a square mile of prairie (Def 3.11) A r.v. Y is said to have a Poisson probability distribution with the parameter λ (i.e., Y P oisson(λ)) if and only if p(y) = λy y! e λ, where y = 0, 1, 2,..., and λ > 0 (λ does not have to be an integer, but Y does). Here, λ (rate)= average of (rare) events that occur in a fixed space, time interval, volume, or any other dimensions (i.e., number of occurrences per that unit of dimension). 14

(Question) Does p(y) in (Def 3.11) satisfy the necessary properties in (Theorem 3.1)? (Hint) Use the following fact : e λ = 1 + λ + λ2 2! + λ3 3! + = y=0 λ y y!. (Theorem 3.11) Let Y be a random variable with a poisson distribution, µ = E(Y ) = λ, σ 2 = V (Y ) = λ. (Example) If Y P oi(λ) and σ 2 = 3, P (Y = 2)? (Example) Suppose Y P oi(λ) so that 3P (Y = 1) = P (Y = 2). Find P (Y = 4)? (Example) The mean of a poisson r.v. Y is µ = 9. Compute P (µ 2σ < Y < µ + 2σ). 15

(Example)The average number of homes sold by the X Realty company is 3 homes per day. What is the probability that exactly 4 homes will be sold tomorrow? (Exercise 3.127) The number of typing errors made by a typist has a poisson distribution with an average of four errors per page. If more than four errors appear on a given page, the typist must retype the whole page. What is the probability that a certain page does not have to be typed? (Example) Suppose that Y is the number of accidents in a 4 hours window, and the number of accidents per hour is 3. Then, what is P (Y = 2)? 16

[Relationship between Binomial distribution and Poisson distribution] Suppose Y is a Binomial r.v. with parameters n (total number of trials) and p (probability of success). For large n and small p such that λ = np, the following approximation can be used : lim P (Y = y) = lim n n ( n y ) p y (1 p) n y = λy y! e λ This approximation is quite accurate if either n 20 and p 0.05 or n 100 and p 0.10. (Example) Y b(100, 0.02) (a) P (Y 3) = (b) Using Poisson distribution, approximate P (Y 3). 17

3.11 Tchebysheff s Theorem How one can approximate the probability that the r.v. Y is within a certain interval? Use Empirical Rule if the probability distribution of Y is approximately bell-shaped. The interval with endpoints, (µ σ, µ + σ) contains approximately 68 % of the measurements. (µ 2σ, µ+2σ) contains approximately 95 % of the measurements. (µ 3σ, µ + 3σ) contains approximately 99.7 % of the measurements. E.G.) suppose that the scores on STAT515 midterm exam have approximately a bell-shaped curve with µ = 80 and σ = 5. Then approximately 68% of the scores are between 75 and 85, approximately 95% of the scores are between 70 and 90, almost all of the scores are between 65 and 95. 18

However, how one can approximate the probability that the r.v. Y is within a certain interval when the shape of its probability distribution is not bell-shaped? (Theorem 4.13)[Tchebysheff s Theorem] Let Y be a r.v. with µ = E(Y ) < and σ 2 = V ar(y ) <. Then, for any k > 0, P ( Y µ < kσ) 1 1 k 2 or P ( Y µ kσ) 1 k 2 Even if the exact distributions are unknown for r.v. of interest, knowledge of the associated means and standard deviations permits us to determine a (meaningful) lower bounds for the probability that the r.v. Y falls in an intervals µ ± kσ. (Example 3.28) The number of customers per day at a sales counter Y, has been observed for a long period of time and found to have mean 20 and standard deviation 2. The probability distribution of Y is not known. What is the probability that tomorrow Y will be greater than 16 but less than 24? 19