BINOMIAL DISTRIBUTION

Similar documents
Lecture 12. Poisson random variables

Common Discrete Distributions

Suppose that you have three coins. Coin A is fair, coin B shows heads with probability 0.6 and coin C shows heads with probability 0.8.

Expectations. Definition Let X be a discrete rv with set of possible values D and pmf p(x). The expected value or mean value of X, denoted by E(X ) or

p. 4-1 Random Variables

Lecture 3. Discrete Random Variables

Relationship between probability set function and random variable - 2 -

Week 6, 9/24/12-9/28/12, Notes: Bernoulli, Binomial, Hypergeometric, and Poisson Random Variables

Guidelines for Solving Probability Problems

CMPSCI 240: Reasoning Under Uncertainty

Discrete Distributions

MATH MW Elementary Probability Course Notes Part I: Models and Counting

Discrete Distributions

Random variables. DS GA 1002 Probability and Statistics for Data Science.

Probability Theory and Statistics (EE/TE 3341) Homework 3 Solutions

Special Mathematics Discrete random variables

Random Variables Example:

Chapter 4 : Discrete Random Variables

Bernoulli Trials, Binomial and Cumulative Distributions

Binomial and Poisson Probability Distributions

Probability, For the Enthusiastic Beginner (Exercises, Version 1, September 2016) David Morin,

Binomial random variable

Lecture 13. Poisson Distribution. Text: A Course in Probability by Weiss 5.5. STAT 225 Introduction to Probability Models February 16, 2014

Lecture 2: Discrete Probability Distributions

Probability and Statistics for Engineers

Topic 3: The Expectation of a Random Variable

4. Suppose that we roll two die and let X be equal to the maximum of the two rolls. Find P (X {1, 3, 5}) and draw the PMF for X.

Introduction to Probability and Statistics Slides 3 Chapter 3

Notes for Math 324, Part 17

Introduction to Statistical Data Analysis Lecture 3: Probability Distributions

Lecture 8 : The Geometric Distribution

Discrete Probability Distribution

Random Variable. Discrete Random Variable. Continuous Random Variable. Discrete Random Variable. Discrete Probability Distribution

Class 26: review for final exam 18.05, Spring 2014

Conditional Probability

Topic 9 Examples of Mass Functions and Densities

MATH 3670 First Midterm February 17, No books or notes. No cellphone or wireless devices. Write clearly and show your work for every answer.

EECS 126 Probability and Random Processes University of California, Berkeley: Fall 2014 Kannan Ramchandran September 23, 2014.

CS 237: Probability in Computing

MATH 250 / SPRING 2011 SAMPLE QUESTIONS / SET 3

EE126: Probability and Random Processes

Question Points Score Total: 137

Brief Review of Probability

STAT509: Discrete Random Variable

Discrete random variables and probability distributions

Topic 3 - Discrete distributions

Discrete Probability Distributions

18.175: Lecture 13 Infinite divisibility and Lévy processes

b. ( ) ( ) ( ) ( ) ( ) 5. Independence: Two events (A & B) are independent if one of the conditions listed below is satisfied; ( ) ( ) ( )

Chapters 3.2 Discrete distributions

ST 371 (V): Families of Discrete Distributions

II. The Binomial Distribution

Lecture 4: Bernoulli Process

ELEG 3143 Probability & Stochastic Process Ch. 2 Discrete Random Variables

Discrete Random Variables

Methods of Mathematics

Discrete distribution. Fitting probability models to frequency data. Hypotheses for! 2 test. ! 2 Goodness-of-fit test

Chapter (4) Discrete Probability Distributions Examples

Lecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes

DISCRETE VARIABLE PROBLEMS ONLY

3.4. The Binomial Probability Distribution

Things to remember when learning probability distributions:

More Discrete Distribu-ons. Keegan Korthauer Department of Sta-s-cs UW Madison

CS 1538: Introduction to Simulation Homework 1

Fundamental Tools - Probability Theory II

Geometric Distribution The characteristics of a geometric experiment are: 1. There are one or more Bernoulli trials with all failures except the last

Outline PMF, CDF and PDF Mean, Variance and Percentiles Some Common Distributions. Week 5 Random Variables and Their Distributions

Probability Distributions.

Some Special Discrete Distributions

Fault-Tolerant Computer System Design ECE 60872/CS 590. Topic 2: Discrete Distributions

4. Discrete Probability Distributions. Introduction & Binomial Distribution

SDS 321: Introduction to Probability and Statistics

STAT 516 Midterm Exam 2 Friday, March 7, 2008

Math/Stat 352 Lecture 8

STAT2201. Analysis of Engineering & Scientific Data. Unit 3

18.175: Lecture 17 Poisson random variables

MA 250 Probability and Statistics. Nazar Khan PUCIT Lecture 15

Probability Density Functions and the Normal Distribution. Quantitative Understanding in Biology, 1.2

4. The Negative Binomial Distribution

Topic 3: The Expectation of a Random Variable

Binomial in the Limit

Math 493 Final Exam December 01

Introduction to Probability, Fall 2009

Random Variables. Definition: A random variable (r.v.) X on the probability space (Ω, F, P) is a mapping

Statistics for Economists. Lectures 3 & 4

1. I had a computer generate the following 19 numbers between 0-1. Were these numbers randomly selected?

Random variable X is a mapping that maps each outcome s in the sample space to a unique real number x, x. X s. Real Line

Discrete Distributions

CS 361: Probability & Statistics

Closed book and notes. 120 minutes. Cover page, five pages of exam. No calculators.

ACCESS TO SCIENCE, ENGINEERING AND AGRICULTURE: MATHEMATICS 2 MATH00040 SEMESTER / Probability

Expected Value - Revisited

SIMPLE RANDOM WALKS: IMPROBABILITY OF PROFITABLE STOPPING

Midterm Exam 1 Solution

Stat 100a, Introduction to Probability.

What is Probability? Probability. Sample Spaces and Events. Simple Event

1: PROBABILITY REVIEW

Distribusi Binomial, Poisson, dan Hipergeometrik

Stats Review Chapter 6. Mary Stangler Center for Academic Success Revised 8/16

Statistics for Managers Using Microsoft Excel/SPSS Chapter 4 Basic Probability And Discrete Probability Distributions

Transcription:

BINOMIAL DISTRIBUTION

The binomial distribution is a particular type of discrete pmf. It describes random variables which satisfy the following conditions: 1 You perform n identical experiments (called trials), where you know n before starting. 2 The random variable counts the number of times a particular event is seen in all of the trials. 3 Each trial is independent of every other. 4 The probability of seeing the event you re interested in (p) is the same for every trial. Binomial distribution

Example I flip a coin 50 times; let X be the number of times I see heads. I inspect 10 products from an assembly line; let Y be the number of products which fail. A family has 4 children; let Z be the number of boys. Binomial distribution

Counterexamples The following are not binomial random variables: 1 I flip a coin until I see heads; let X be the number of flips. (n is not fixed.) 2 I deal 7 cards; let Y be the number of spades in my hand. (The trials are not independent.) 3 I play in a 5-round tournament; let Z be the number of matches I win. (The probability of winning is different for each opponent.) Binomial distribution

Regardless of the details of the specific case, in terms of probability all binomial random variables behave in the same way, determined by two parameters: n and p. Without loss of generality, we say each experiment has only two possibilities: success (you see the outcome you re interested in) or failure (you don t.) This is the first distribution family we see. There are many more, which describe different kinds of random variables. For each distribution family, we will derive formulas for the pmf, expected value, and variance. Whenever you recognize that a random variable belongs to one of these families, you can use these formulas directly. Binomial distribution

Let s say I know n and p. What is the pmf of X? Step one: what are the possible values X can take? X {0, 1,..., n} Step two: what is the probability that X takes any of these values? Binomial distribution

Example I inspect 10 products from an assembly line; let Y be the number of products which fail. The probability any product passes is 0.9. What is the probability that Y = 2? The probability of the first two products failing and the last eight products passing is (0.1) 2 (0.9) 8. However, I( don t ) care about the order in which the products pass and fail. 10 There are ways in which I can see two products pass and eight fail. 2 Each of these( orderings ) has the same probability. Therefore 10 P(Y = 2) = (0.1) 2 (0.9) 8 = 0.19 2 Binomial distribution

For any binomial random variable X we have P(X = k) = ( ) n p k (1 p) n k k for any k {0, 1,..., n}. Binomial distribution

More Examples I ( flip a coin 50 times; the probability of seeing 20 heads is 50 20) (1/2) 20 (1/2) 30 = 0.042 The probability of winning a casino game is 0.4. If I play 10 times, the probability I win just once is ( ) 10 1 (0.4) 1 (0.6) 9 = 0.040 A ( family has 4 children; the probability of having 2 boys is 4 ) 2 (1/2) 2 (1/2) 2 = 0.375 Binomial distribution

The Bernoulli distribution is the special case of the binomial distribution when n = 1. Binomial distribution

HYPERGEOMETRIC DISTRIBUTION

Remember the four properties of the binomial distribution: 1 You perform n identical experiments (called trials), where you know n before starting. 2 The random variable counts the number of times a particular event is seen in all of the trials. 3 Each trial is independent of every other. 4 The probability of seeing the event you re interested in (p) is the same for every trial. The hypergeometric distribution changes properties 3 and 4. Hypergeometric Distribution

Rather than every trial being completely independent, imagine that we re sampling from a population without replacement. One of our binomial counterexamples was how many spades will I see in a seven-card hand? Each card you are dealt changes the probability of seeing spades for the rest of your hand. Hypergeometric Distribution

The properties of the hypergeometric distribution can be stated as follows: 1 I sample n objects from a finite population which has b successes and r failures. 2 The random variable counts the number of times a particular event is seen in all of the trials. 3 Each sample of k objects is equally likely to be chosen. This is approximately a binomial distribution when b, r, and k are large. Hypergeometric Distribution

What is the probability of being dealt 3 aces in a hand of 5 cards? In terms of the hypergeometric distribution, we have b = 4 (successes in population), r = 48 (failures in population), and k = 5 (sample size). Hypergeometric Distribution

Since each hand of 5 cards is equally likely, we can use a counting argument like that used to derive properties of the binomial distribution. The number of 5-card hands is P(A) = N(A) N ( ) 52 = 2598960. 5 Each hand we are interested in has 3 aces, and 2 non-aces. ( ) 4 The number of ways to choose the 3 aces is = 4. (Why?) 3 ( ) 48 The number of ways to choose the 2 non-aces is = 1128. 2 By the fundamental principle of counting, N(A) = 4 2256 = 4512. Therefore, the probability of seeing three aces is 4512/2598960 = 0.00174, or roughly 1 in 576. Hypergeometric Distribution

We can generalize this procedure to find the probability that X = x when X is a hypergeometric random variable. ( ) b + r The number of different samples we could have seen is. k Each hand we are interested in has x successes, and k x failures. ( ) b The number of ways to choose the x successes is. x ( ) r The number of ways to choose the k x failures is. k x By the fundamental ( )( principle ) of counting, the number of samples with x b r successes is. x k x ( )( ) b r P(X = x) = x k x ( ) b + r k Hypergeometric Distribution

Example Out of the 46 students in class, 10 will come to my office hours this week. Assuming that there are 20 students in architectural engineering, and that the students who come to my office hours are a random sample, what is the probability that four of them are from architectural? Define a student from architectural engineering as a success. Then b = 20, r = 26, k = 10, and x = 5. ( )( ) 20 26 P(X = 4) = 4 6 ( ) = 0.27 46 10 Hypergeometric Distribution

NEGATIVE BINOMIAL

Remember the four properties of the binomial distribution: 1 You perform n identical experiments (called trials), where you know n before starting. 2 The random variable counts the number of times a particular event is seen in all of the trials. 3 Each trial is independent of every other. 4 The probability of seeing the event you re interested in (p) is the same for every trial. The negative binomial distribution changes properties 1 and 2. Negative Binomial

Rather than counting the number of successes in a predetermined number of trials, we want to know how many trials it will take before we see a predetermined number of successes. Specifically, a negative binomial random variable counts the number of trials until the m-th success (including the last one), where m is given. I flip a coin until I see 3 heads. How many times do I have to flip the coin? Negative Binomial

You are interviewing students to hire 3 interns. Assume that the probability each interviewee is an acceptable hire is 30%. What is the probability I will have to interview 8 students in total? In terms of the negative binomial distribution, we have m = 3 (the number of successes before I stop) and p = 0.3 (probability that each interviewee is an acceptable hire). Negative Binomial

I know that the last student I interview was a success. Therefore, of the preceding students, there must be 5 who were rejected, and 2 who were hired, for a total of 8 interviews. The probability that the 8th interview was a success is 0.3. The probability of having 2 hires from the first 7 is binomial, with n = 7, p = 0.3, and x = 2. Therefore, the probability is (( ) 7 )(0.3) 2 (0.7) 5 (0.3). 2 Negative Binomial

For any negative binomial random variable, we have P(X = k) = ( ) k 1 p m (1 p) k m m 1 for k {m, m + 1,...} and 0 otherwise. Negative Binomial

The negative binomial distribution is also known as the Pascal distribution. The geometric distribution is the special case when m = 1. Some books define the negative binomial distribution slightly differently, to count the number of failures before the m-th success rather than the number of trials. The concepts are the same, but the formulas will be slightly different (the random variables will differ by m). Negative Binomial

POISSON DISTRIBUTION

One common scenario involves random occurrences over time. Examples: How many customers arrive at a checkout line in an hour? How many vehicles drive past a traffic signal during 15 minutes? How many goals will be scored in a soccer match? None of the distributions we ve seen so far represent this situation. Poisson Distribution Motivation

Enter the Poisson distribution Poisson Distribution Motivation

The Poisson distribution is used to count the number of times an event occurs, assuming 1 Occurrences of the event are independent of each other. 2 More than one of these events cannot occur simultaneously. Poisson Distribution Definition

These two assumptions hold true for the scenarios we saw earlier. How many transactions will a cashier handle in an hour? How many vehicles drive past a traffic signal during 15 minutes? How many goals will be scored in a soccer match? Poisson Distribution Definition

Here are a few counterexamples which do not satisfy the Poisson assumptions: Why? How many buses will pass the bus stop in an hour? How many people arrive at a checkout line in an hour? Poisson Distribution Definition

To derive the Poisson pmf, assume we know the average number of occurrences λ over the time period, and divide the time period into n equal intervals. Then P(k) can be found using a binomial distribution. The Poisson pmf is found by taking the limit as n. Poisson Distribution Definition

If there are n equal intervals, and the average rate is λ, this means p = λ/n. ( ) n For the binomial distribution P(k) = p k (1 p) n k. k lim P(k) = lim n n ( ) n p k (1 p) n k = lim k n k! k!(n k)! ( = λk n(n 1) (n k + 1) lim 1 λ ) n k k! n n(n) (n) n ( = λx lim 1 λ ) n k! n n = e λ λ k k! ( ) λ k ( 1 λ ) n k n n Poisson Distribution Definition

Assume that the number of students who stop by my office hours is a Poisson random variable with λ = 2. What is the probability that no students stop by my office hours? The probability that between 3 and 5 students stop by? P(X = 0) = e 2 2 0 0! = 0.135 P(X = 3) + P(X = 4) + P(X = 5) = e 2 2 3 3! + e 2 2 4 4! + e 2 5 0 5! = 0.307 Poisson Distribution Definition

The number of callers to a technical support line is a Poisson random variable. On average, there are 2 calls per hour. What is the standard deviation of the number of calls in an eight-hour shift? What is the probability that there will be at least 2 calls during an eight-hour shift? For the eight-hour shift, the average number of occurrences is 16, so λ = 16. V [X ] = 16 = 4 P(X 2) = 1 P(X = 0) P(X = 1) = 1 e 16 16 0 0! + e 16 16 1 1! = 0.999998 Poisson Distribution Definition

For the Poisson PMF to be valid, P(X = k) 0 for all k, and P(X = k) = 1. Are these true? k Poisson Distribution Definition