Hypergeometric, Poisson & Joint Distributions

Similar documents
Bernoulli Trials and Binomial Distribution

Bernoulli Trials, Binomial and Cumulative Distributions

Bernoulli Trials and Binomial Distribution

STAT/MATH 395 PROBABILITY II

Conditional distributions (discrete case)

Bivariate Distributions

Outline PMF, CDF and PDF Mean, Variance and Percentiles Some Common Distributions. Week 5 Random Variables and Their Distributions

Covariance and Correlation

Expected Values, Exponential and Gamma Distributions

Math Key Homework 3 (Chapter 4)

Expected Values, Exponential and Gamma Distributions

Lecture 13. Poisson Distribution. Text: A Course in Probability by Weiss 5.5. STAT 225 Introduction to Probability Models February 16, 2014

f X, Y (x, y)dx (x), where f(x,y) is the joint pdf of X and Y. (x) dx

Jointly Distributed Random Variables

STAT Chapter 5 Continuous Distributions

Notes for Math 324, Part 19

STOR Lecture 16. Properties of Expectation - I

Gamma and Normal Distribuions

ENGG2430A-Homework 2

HW5 Solutions. (a) (8 pts.) Show that if two random variables X and Y are independent, then E[XY ] = E[X]E[Y ] xy p X,Y (x, y)

STAT509: Discrete Random Variable

STA 256: Statistics and Probability I

STAT/MA 416 Answers Homework 6 November 15, 2007 Solutions by Mark Daniel Ward PROBLEMS

Bayes Theorem and Hypergeometric Distribution

Chapter 3: Discrete Random Variable

Chapter 5 continued. Chapter 5 sections

Statistics 427: Sample Final Exam

Bivariate distributions

HW on Ch Let X be a discrete random variable with V (X) = 8.6, then V (3X+5.6) is. V (3X + 5.6) = 3 2 V (X) = 9(8.6) = 77.4.

More than one variable

Covariance. Lecture 20: Covariance / Correlation & General Bivariate Normal. Covariance, cont. Properties of Covariance

Chapter 4 continued. Chapter 4 sections

MA 575 Linear Models: Cedric E. Ginestet, Boston University Revision: Probability and Linear Algebra Week 1, Lecture 2

ECSE B Solutions to Assignment 8 Fall 2008

Exponential, Gamma and Normal Distribuions

Ch. 5 Joint Probability Distributions and Random Samples

Statistics for Economists Lectures 6 & 7. Asrat Temesgen Stockholm University

Midterm Exam 1 Solution

Introduction to Computational Finance and Financial Econometrics Probability Review - Part 2

CHAPTER 4 MATHEMATICAL EXPECTATION. 4.1 Mean of a Random Variable

Introduction to Statistical Data Analysis Lecture 3: Probability Distributions

Confidence Intervals for Two Means

The Binomial distribution. Probability theory 2. Example. The Binomial distribution

Chapter 5. Chapter 5 sections

Lecture 2: Repetition of probability theory and statistics

Test 1 Review. Review. Cathy Poliak, Ph.D. Office in Fleming 11c (Department Reveiw of Mathematics University of Houston Exam 1)

3. Probability and Statistics

Homework 4 Solution, due July 23

Lecture 4: Random Variables and Distributions

Probability and Statistics Notes

The Normal Distribuions

1.1 Review of Probability Theory

Class 8 Review Problems solutions, 18.05, Spring 2014

LIST OF FORMULAS FOR STK1100 AND STK1110

Random Variable And Probability Distribution. Is defined as a real valued function defined on the sample space S. We denote it as X, Y, Z,

3 Multiple Discrete Random Variables

Let X and Y denote two random variables. The joint distribution of these random

t x 1 e t dt, and simplify the answer when possible (for example, when r is a positive even number). In particular, confirm that EX 4 = 3.

Math Spring Practice for the final Exam.

Chapter 5 Class Notes

BINOMIAL DISTRIBUTION

Joint probability distributions: Discrete Variables. Two Discrete Random Variables. Example 1. Example 1

2. Suppose (X, Y ) is a pair of random variables uniformly distributed over the triangle with vertices (0, 0), (2, 0), (2, 1).

CMPSCI 240: Reasoning Under Uncertainty

Chapters 3.2 Discrete distributions

Math 510 midterm 3 answers

1 Basic continuous random variable problems

Problem Solving. Correlation and Covariance. Yi Lu. Problem Solving. Yi Lu ECE 313 2/51

Expectation of Random Variables

Mathematical Statistics 1 Math A 6330

Random Variables. P(x) = P[X(e)] = P(e). (1)

CMPSCI 240: Reasoning Under Uncertainty

Lecture 16 : Independence, Covariance and Correlation of Discrete Random Variables

MATHEMATICS 154, SPRING 2009 PROBABILITY THEORY Outline #11 (Tail-Sum Theorem, Conditional distribution and expectation)

Massachusetts Institute of Technology Department of Electrical Engineering & Computer Science 6.041/6.431: Probabilistic Systems Analysis

STAT 430/510: Lecture 16

Random Variables. Cumulative Distribution Function (CDF) Amappingthattransformstheeventstotherealline.

Functions of two random variables. Conditional pairs

Review of Probability. CS1538: Introduction to Simulations

Outline Properties of Covariance Quantifying Dependence Models for Joint Distributions Lab 4. Week 8 Jointly Distributed Random Variables Part II

Appendix A : Introduction to Probability and stochastic processes

Random Variables. Lecture 6: E(X ), Var(X ), & Cov(X, Y ) Random Variables - Vocabulary. Random Variables, cont.

MAS113 Introduction to Probability and Statistics. Proofs of theorems

SUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416)

STAT 515 MIDTERM 2 EXAM November 14, 2018

Bivariate Distributions. Discrete Bivariate Distribution Example

Closed book and notes. 60 minutes. Cover page and four pages of exam. No calculators.

MATH 151, FINAL EXAM Winter Quarter, 21 March, 2014

Statistics for Economists. Lectures 3 & 4

Discrete Distributions

Algorithms for Uncertainty Quantification

Jointly Distributed Variables

Homework 10 (due December 2, 2009)

Tom Salisbury

Discrete Probability Distribution

Chapter 3 Discrete Random Variables

Expectations. Definition Let X be a discrete rv with set of possible values D and pmf p(x). The expected value or mean value of X, denoted by E(X ) or

Sample Problems for the Final Exam

Continuous Random Variables

Transcription:

Hypergeometric, Poisson & Joint Distributions Sec 4.7-4.9 Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Department of Mathematics University of Houston Lecture 6-3339 Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University of Houston Lecture 6) - 3339 1 / 30

Outline 1 Hypergeometric Distribution 2 Poisson Distribution 3 Joint Distributions Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University of Houston Lecture 6) - 3339 2 / 30

Beginning Example Each of 12 refrigerators of a certain type has been returned to a distributor because of an audible, high-pitched, oscillating noise when the refrigerator is running. Suppose that 7 of these refrigerators have a defective compressor and the other 5 have less serious problems. The technition looks at 6 refrigerators, what is the probability that exactly 5 have a defective compressor? Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University of Houston Lecture 6) - 3339 3 / 30

Conditions for a Hypergeometric Distribution 1. The population or set to be sampled consists of N individuals, objects or elements (a finite population). 2. Each individual can be characterized as a "success" or "failure." There are m successes in the population, and n failures in the population. Notice: m + n = N. 3. A sample size of k individuals is selected without replacement in such a way that each subset of size k is equally likely to be chosen. The parameters of a hypergeometric distribution is m, n, k. We write X Hyper(m, n, k). The probability mass function for a hypergeometric is: ( m )( n ) x f X (x) = P(X = x) = k x ( m+n k ) Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University of Houston Lecture 6) - 3339 4 / 30

Conditions for a Hypergeometric Distribution 1. The population or set to be sampled consists of N individuals, objects or elements (a finite population). 2. Each individual can be characterized as a "success" or "failure." There are m successes in the population, and n failures in the population. Notice: m + n = N. 3. A sample size of k individuals is selected without replacement in such a way that each subset of size k is equally likely to be chosen. The parameters of a hypergeometric distribution is m, n, k. We write X Hyper(m, n, k). The probability mass function for a hypergeometric is: ( m )( n ) x f X (x) = P(X = x) = k x ( m+n k ) Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University of Houston Lecture 6) - 3339 4 / 30

Using R R commands:p(x = x) = dhyper(x,m,n,k) and P(X x) = phyper(x, m, n, k) Example going back to the refrigerator example, m = 7, n = 5, k = 6. P(X = 5) > dhyper(5,7,5,6) [1] 0.1136364 P(X 4) > phyper(4,7,5,6) [1] 0.8787879 Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University of Houston Lecture 6) - 3339 5 / 30

Mean and Variance of a Hypergeometric Distribution Let Y have a hypergeometric distribution with parameter, m, n, and k. The mean of Y is: µ Y = E(Y ) = k ( ) m = kp. m + n The variance of Y is: ( σy 2 = var(y ) = kp(1 p) 1 k 1 ). m + n 1 1 k 1 m+n 1 is called the finite population correction factor. As, the population increases, this factor will get closer to 1. Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University of Houston Lecture 6) - 3339 6 / 30

Digial Cameras A certain type of digital camera comes in either a 3-megapixel version or a 4-megapixel version. A camera store has received a shipment of 15 of these cameras, of which 6 have 3-megapixel resolution. Suppose that 5 of these cameras are randomly selected to be stored behind the counter; the other 10 are placed in a storeroom. Let X = the number of 3-megapixel cameras among the 5 selected for behind the counter storage. 1. What is the probability that exactly 2 of the 3-megapixel cameras are stored behind the counter? 2. What is the probability that at least one of the 3-megapixel cameras are stored behind the counter? 3. Calculate the mean and standard deviation of X. Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University of Houston Lecture 6) - 3339 7 / 30

Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University of Houston Lecture 6) - 3339 8 / 30

Example Suppose that the average number of emails received by a particular student at UH is five emails per hour. We want to know what is the probability that a particular student will get X number of emails in any given hour. Suppose we want to know P(X = 3), the probability of getting exactly 3 emails in one hour. What about in two hours? Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University of Houston Lecture 6) - 3339 9 / 30

The Poisson Distribution The Poisson distribution is appropriate under the following conditions. 1. Let X be the number of successes occurring per unit of measure. X = 0, 1, 2, 3,.... 2. Let µ be the mean number of successes occurring per unit of measure. 3. The number of successes that occur in two nonoverlapping units of measure are independent. 4. The probability that success will occur in a unit of measure is the same for all unites of equal size and is proportional to the size of the unit. 5. The probability that more than one event occurs in a unit of measure is negligible for very small-sized units. In other words, the events occur one at a time. Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University oflecture Houston 6 )- 3339 10 / 30

The Probability Function of the Poisson Distribution A random variable X with nonnegative integer values has a Poisson distribution if its frequency function is: µ µx f (x) = P(X = x) = e x! for x = 0, 1, 2,..., where µ > 0 is a constant. If X has a Poisson distribution with parameter µ, we can write X Pois(µ). Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University oflecture Houston 6 )- 3339 11 / 30

The Mean and Variance of the Poisson Distribution Let X Pois(µ) The mean of X is µ per unit of measure. By the conditions of the Poisson distribution. The variance of X is also µ per unit of measure. The standard deviation of X is µ. Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University oflecture Houston 6 )- 3339 12 / 30

Example The number of people arriving for treatment at an emergency room can be modeled by a Poisson process with a mean of five people per hour. 1. What is the probability that exactly four arrivals occur at a particular hour? 2. What is the probability that at least four people arrive during a particular hour? 3. How many people do you expect to arrive during a 45-min period? Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University oflecture Houston 6 )- 3339 13 / 30

Using R to Compute Probabilities of the Poisson Distribution R commands:p(x = x) = dpois(x,µ) and P(X x) = ppois(x, µ) P(X =4) > dpois(4,5) [1] 0.1754674 P(X 4) = 1 P(X 3) > 1-ppois(3,5) [1] 0.7349741 Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University oflecture Houston 6 )- 3339 14 / 30

Type of Probabiltiy Distributions For the following information indicate the type of distribution. a. Binomial b. Poisson c.hypergoemetric d. Neither 1. A company makes sports bikes. 90% pass final inspection (and 10% fail and need to be fixed). We randomly select 10 bikes and want to know what is the probability that 7 will pass. 2. A company makes sports bikes. 90% pass final inspection (and 10% fail and need to be fixed). What is the probability that the 10 th bike will not pass final inspection? 3. A small company made 100 bikes in a month. 10 of these bikes will need to be fixed. Suppose we select 4 bikes, what is the probability that at least one needs to be fixed? 4. A small company makes bicycles. The average amount that this company makes in a month is 100 bikes. What is the probability that the company will make less than 30 bikes in a week? Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University oflecture Houston 6 )- 3339 15 / 30

Begining Example Suppose two random variables X and Y have a joint function f (x, y) = 2x y 5 for X = 0.75 and 1 and Y = 0 and 1. What are the possible values of f (x, y)? Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University oflecture Houston 6 )- 3339 16 / 30

Joint Probability Distribution If X and Y are two discrete random variables, the probability distribution for their simultaneous occurrence can be represented by a function with values f (x, y) for any pair of values (x, y) within the range of the random variables X and Y. It is customary to refer this function as the joint probability function of X and Y. When X and Y are discrete the probability function is: f (x, y) = P(X = x and Y = y) The marginal frequency functions of X and Y are simply their individual frequency functions. f (x) = P(X = x) f (y) = P(Y = y) These functions is best represented in a joint probability table (contingency table). Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University oflecture Houston 6 )- 3339 17 / 30

Theorem 4.7 Let X and Y have the joint frequency function f (x, y). Then 1. f Y (y) = x f (x, y) for all y. 2. f X (x) = y f (x, y) for all x. 3. f Y X (y x) = f (x,y) f (x) function. if f X (x) > 0. This is the conditional frequency 4. X and Y are independent if and only if f (x, y) = f X (x)f Y (y) for all x, y. 5. X Y f (x, y) = 1. Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University oflecture Houston 6 )- 3339 18 / 30

Popper 07 Questions Suppose that the following table represents the joint pmf of X and Y. Y /X 1 2 1 2/k 3/k 2 3/k 4/k 3 4/k 5/k 1. What should be the vaule of k? 2. Determine P(X = 2, Y 2) 3. Determine P(X = 2 Y = 2) a) 6 b) 21 c) 5 d) 100 a) 4 7 b) 3 7 c) 1 3 d) 4 21 a) 4 7 b) 3 7 c) 1 3 d) 4 21 Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University oflecture Houston 6 )- 3339 19 / 30

Rule 2 for Means If X and Y are two different random variables, then the mean of the sums of the pairs of the random variable is the same as the sum of their means: E[X + Y ] = E[X] + E[Y ]. This is called the addition rule for means. The mean of the difference of the pairs of the random variable is the same as the difference of their means: E[X Y ] = E[X] E[Y ]. Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University oflecture Houston 6 )- 3339 20 / 30

Rule 2 for Variances If X and Y are independent random variables and σ 2 X+Y σ 2 X Y = Var[X + Y ] = Var[X] + Var[Y ]. = Var[X Y ] = Var[X] + Var[Y ]. Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University oflecture Houston 6 )- 3339 21 / 30

Example of Rule 2 Tamara and Derek are sales associates in a large electronics and appliance store. The following table shows their mean and standard deviation of daily sales. Assume that daily sales among the sales associates are independent. Sales associate Mean Standard deviation Tamara X E[X] = $1100 σ X = $100 Derek Y E[Y ] = $1000 σ Y = $80 1. Determine the mean of the total of Tamara s and Derek s daily sales, E[X + Y ]. 2. Determine the variance of the total of Tamara s and Derek s daily sales, σ 2 (X+Y ). 3. Determine the standard deviation of the total of Tamara s and Derek s daily sales, σ (X+Y ). Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University oflecture Houston 6 )- 3339 22 / 30

Example Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University oflecture Houston 6 )- 3339 23 / 30

Expected Values Let X and Y be two random variables, then: E(X + Y ) = E(X) + E(Y ) E(XY ) = x 1 y 1 p 11 + x 1 y 1 p 12 +... + x n y n p nn For more than two random variables X 1, X 2, X 3,, X n, E(X 1 + X 2 + + X n ) = E(X 1 ) + E(X 2 ) + + E(X n ) Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University oflecture Houston 6 )- 3339 24 / 30

Covariance of X and Y Let X and Y be jointly distributed random variables with respective means µ x and µ y and standard deviation σ x, σ y. The covariance of X and Y is cov(x, Y ) = E((x µ x )(Y µ y )). or an easier calculation: The correlation of X and Y is; cov(x, Y ) = E(XY ) E(X)E(Y ). cor(x, Y ) = cov(x, Y ) σ x σ y. Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University oflecture Houston 6 )- 3339 25 / 30

Properties of Covariance 1. cov(x, Y ) = cov(y, X) 2. cov(x, X) = var(x) 3. If X, Y, and Z are jointly distributed and a and b are constants cov(x, ay + bz ) = a[cov(x, Y )] + b[cov(x, Z )]. 4. If X and Y are jointly distributed, var(x + Y ) = var(x) + var(y ) + 2cor(X, Y )sd(x)sd(y ) var(x Y ) = var(x) + var(y ) 2cor(X, Y )sd(x)sd(y ) 5. If X and Y are independent, cov(x, Y ) = 0. 6. If jointly distributed random variables X 1, X 2,, X n are pairwise uncorrelated, then var(x 1 + X 2 + + X n ) = var(x 1 ) + var(x 2 ) + + var(x n ) Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University oflecture Houston 6 )- 3339 26 / 30

Example of a Joint Probability Distribution The following table is a joint probability table of X = number of sports involved and Y = age of high school students. 1. Determine E(X). 2. Determine E(Y ). 3. Determine E(XY ). 4. Determine cov(x, Y ). X = Number of sports involved Y = Age 0 1 2 3 14 0.03 0.09 0.06 0.02 15 0.015 0.045 0.03 0.01 16 0.045 0.135 0.09 0.03 17 0.03 0.09 0.06 0.02 18 0.03 0.09 0.06 0.02 Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University oflecture Houston 6 )- 3339 27 / 30

R code I put the values in an Excel.csv file named "jointprob." > library(readr) > jointprob <- read_csv("f:/lectures uh/math 3339/jointprob > View(jointprob) > sum(jointprob$x*jointprob$y*jointprob$pxy) [1] 21.735 > 1.35*16.1 [1] 21.735 > sum(jointprob$x*jointprob$pxy) [1] 1.35 > sum(jointprob$y*jointprob$pxy) [1] 16.1 > sum(jointprob$x*jointprob$y*jointprob$pxy) [1] 21.735 Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University oflecture Houston 6 )- 3339 28 / 30

Example 2 A class has 10 mathematics majors, 6 computer science majors and 4 statistics majors. A committee of two is selected at random to work an a problem. Let X be the number of mathematics majors and let Y be the number of computer science majors chosen. 1. Determine all possible (X, Y ) pairs for randomly selecting two students. 2. Determine f X,Y (0, 0), that is the probability that the two pick are neither a mathematics major nor a computer science major. 3. Determine f X (0), that is the probability that we do not pick any mathematics majors. Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University oflecture Houston 6 )- 3339 29 / 30

Joint Probability Table Y /X 0 1 2 Total 0 1 2 Total 1. Determine E(X). 2. Determine E(Y ). 3. Determine E(XY ). 4. Determine cov(x, Y ). 5. Determine cor(x, Y ). Cathy Poliak, Ph.D. cathy@math.uh.edu Office in Fleming 11c Sec (Department 4.7-4.9 of Mathematics University oflecture Houston 6 )- 3339 30 / 30