Math 151. Rumbos Spring Solutions to Review Problems for Exam 3

Similar documents
Math 151. Rumbos Fall Solutions to Review Problems for Exam 2. Pr(X = 1) = ) = Pr(X = 2) = Pr(X = 3) = p X. (k) =

Math 151. Rumbos Spring Solutions to Review Problems for Exam 1

Math 151. Rumbos Spring Solutions to Review Problems for Exam 2

Math 151. Rumbos Fall Solutions to Review Problems for Final Exam

Twelfth Problem Assignment

2. Suppose (X, Y ) is a pair of random variables uniformly distributed over the triangle with vertices (0, 0), (2, 0), (2, 1).

Mathematical Statistics 1 Math A 6330

Fundamental Tools - Probability Theory IV

CS145: Probability & Computing

MATH Solutions to Probability Exercises

Limiting Distributions

Computer Science, Informatik 4 Communication and Distributed Systems. Simulation. Discrete-Event System Simulation. Dr.

MATH 250 / SPRING 2011 SAMPLE QUESTIONS / SET 3

Mathematical statistics

Math 151. Rumbos Fall Solutions to Review Problems for Final Exam

Final Solutions Fri, June 8

ECE 302 Division 2 Exam 2 Solutions, 11/4/2009.

Discrete Mathematics and Probability Theory Fall 2011 Rao Midterm 2 Solutions

Closed book and notes. 60 minutes. Cover page and four pages of exam. No calculators.

ECE302 Exam 2 Version A April 21, You must show ALL of your work for full credit. Please leave fractions as fractions, but simplify them, etc.

More on Distribution Function

Lecture 13. Poisson Distribution. Text: A Course in Probability by Weiss 5.5. STAT 225 Introduction to Probability Models February 16, 2014

Math 151. Rumbos Spring Solutions to Assignment #15

Homework 10 (due December 2, 2009)

Mathematical statistics

Quick Tour of Basic Probability Theory and Linear Algebra

IE 230 Probability & Statistics in Engineering I. Closed book and notes. 60 minutes.

STAT/MATH 395 PROBABILITY II

15 Discrete Distributions

Probability Review. Yutian Li. January 18, Stanford University. Yutian Li (Stanford University) Probability Review January 18, / 27

STAT/MATH 395 A - PROBABILITY II UW Winter Quarter Moment functions. x r p X (x) (1) E[X r ] = x r f X (x) dx (2) (x E[X]) r p X (x) (3)

Week 12-13: Discrete Probability

STAT/MA 416 Midterm Exam 2 Thursday, October 18, Circle the section you are enrolled in:

1. Let X be a random variable with probability density function. 1 x < f(x) = 0 otherwise

3 Continuous Random Variables

STAT 6385 Survey of Nonparametric Statistics. Order Statistics, EDF and Censoring

Exercises in Probability Theory Paul Jung MA 485/585-1C Fall 2015 based on material of Nikolai Chernov

Distributions of Functions of Random Variables. 5.1 Functions of One Random Variable

Applied Statistics I

A Probability Primer. A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes.

Limiting Distributions

Math 447. Introduction to Probability and Statistics I. Fall 1998.

n(1 p i ) n 1 p i = 1 3 i=1 E(X i p = p i )P(p = p i ) = 1 3 p i = n 3 (p 1 + p 2 + p 3 ). p i i=1 P(X i = 1 p = p i )P(p = p i ) = p1+p2+p3

Bivariate distributions

Statistics for Economists. Lectures 3 & 4

Chapter 3 Single Random Variables and Probability Distributions (Part 1)

Chapter 6 Expectation and Conditional Expectation. Lectures Definition 6.1. Two random variables defined on a probability space are said to be

Math Spring Practice for the final Exam.

MAS113 Introduction to Probability and Statistics

STT 441 Final Exam Fall 2013

Lecture 2: Discrete Probability Distributions

M378K In-Class Assignment #1

Review. December 4 th, Review

by Dimitri P. Bertsekas and John N. Tsitsiklis Last updated: November 29, 2002

Probability Models. 4. What is the definition of the expectation of a discrete random variable?

Section 9.1. Expected Values of Sums

APPM/MATH 4/5520 Solutions to Exam I Review Problems. f X 1,X 2. 2e x 1 x 2. = x 2

Introduction to Probability, Fall 2013

Bell-shaped curves, variance

Fundamental Tools - Probability Theory II

Math 493 Final Exam December 01

1.1 Review of Probability Theory

MCS Introduction to Probability and Statistics FIRST MIDTERM

Expectation of Random Variables

MATH 151, FINAL EXAM Winter Quarter, 21 March, 2014

1 Exercises for lecture 1

1 Random Variable: Topics

n px p x (1 p) n x. p x n(n 1)... (n x + 1) x!

STAT 516 Midterm Exam 2 Friday, March 7, 2008

FINAL EXAM: 3:30-5:30pm

Chapter 5. Chapter 5 sections

HW Solution 4 Not Due

ACM 116: Lectures 3 4

Chapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University

Math 3215 Intro. Probability & Statistics Summer 14. Homework 5: Due 7/3/14

Stochastic Models in Computer Science A Tutorial

Definition: A random variable X is a real valued function that maps a sample space S into the space of real numbers R. X : S R

Chapter 5. Statistical Models in Simulations 5.1. Prof. Dr. Mesut Güneş Ch. 5 Statistical Models in Simulations

SDS 321: Introduction to Probability and Statistics

Relationship between probability set function and random variable - 2 -

Midterm #1. Lecture 10: Joint Distributions and the Law of Large Numbers. Joint Distributions - Example, cont. Joint Distributions - Example

STAT2201. Analysis of Engineering & Scientific Data. Unit 3

Probability and Statistics Notes

Introduction to Probability

Joint Probability Distributions and Random Samples (Devore Chapter Five)

1 Basic continuous random variable problems

Continuous Expectation and Variance, the Law of Large Numbers, and the Central Limit Theorem Spring 2014

This does not cover everything on the final. Look at the posted practice problems for other topics.

Statistics STAT:5100 (22S:193), Fall Sample Final Exam B

HW1 (due 10/6/05): (from textbook) 1.2.3, 1.2.9, , , (extra credit) A fashionable country club has 100 members, 30 of whom are

ELEG 3143 Probability & Stochastic Process Ch. 2 Discrete Random Variables

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

(Practice Version) Midterm Exam 2

1 Presessional Probability

Suppose that you have three coins. Coin A is fair, coin B shows heads with probability 0.6 and coin C shows heads with probability 0.8.

Analysis of Engineering and Scientific Data. Semester

X = X X n, + X 2

Probability reminders

(y 1, y 2 ) = 12 y3 1e y 1 y 2 /2, y 1 > 0, y 2 > 0 0, otherwise.

Probability Distributions

Transcription:

Math 151. Rumbos Spring 2014 1 Solutions to Review Problems for Exam 3 1. Suppose that a book with n pages contains on average λ misprints per page. What is the probability that there will be at least m pages which contain more than k missprints? Solution: Let Y denote the number of misprints in one page. Then, we may assume that Y follows a Poisson(λ) distribution; so that Pr[Y = r] = λr r! e λ, for r = 0, 1, 2,... Thus, the probability that there will be more than k missprints in a given page is p = Pr(Y > k) so that = 1 Pr(Y k), p = 1 k r=0 λ r r! e λ. (1) Next, let X denote the number of the pages out of the n that contain more than k missprints. Then, X Binomial(n, p), where p is as given in (1). Then the probability that there will be at least m pages which contain more than k missprints is Pr[X m] = n l=m ( ) n p l (1 p) n l, l where p = 1 k r=0 λ r r! e λ. 2. Suppose that the total number of items produced by a certain machine has a Poisson distribution with mean λ, all items are produced independently of one another, and the probability that any given item produced by the machine will be defective is p. Let X denote the number of defective items produced by the machine.

Math 151. Rumbos Spring 2014 2 (a) Determine the marginal distribution of the number of defective items, X. Solution: Let N denote the number of items produced by the machine. Then, N Poisson(λ), (2) so that Pr[N = n] = λn n! e λ, for n = 0, 1, 2,... Now, since all items are produced independently of one another, and the probability that any given item produced by the machine will be defective is p, X has a conditional distribution (conditioned on N = n) that is Binomial(n, p); thus, ( ) n p k (1 p) n k, for k = 0, 1, 2,..., n; k Pr[X = k N = n] = 0 elsewhere. (3) Then, Pr[X = k] = Pr[X = k, N = n] n=0 = Pr[N = n] Pr[X = k N = n], n=0 where Pr[X = k N = n] = 0 for n < k, so that, using (2) and (3), Pr[X = k] = n=k = e λ k! λ n n! e λ p k n=k ( ) n p k (1 p) n k k λ n 1 (n k)! (1 p)n k. Next, make the change of variables l = n k in the last summation in (4) to get (4) Pr[X = k] = e λ k! p k l=0 λ l+k 1 l! (1 p)l,

Math 151. Rumbos Spring 2014 3 so that Pr[X = k] = (λp)k k! = (λp)k k! e λ l=0 e λ e λ(1 p) 1 [λ(1 p)]l l! which shows that = (λp)k k! e λp, X Poisson(λp). (5) (b) Let Y denote the number of non defective items produced by the machine. Show that X and Y are independent random variables. Solution: Similar calculations to those leading to (5) show that Y Poisson(λ(1 p)), (6) since the probability of an item coming out non defective is 1 p. Next, observe that Y = N X and compute the joint probability Pr[X = k, Y = l] = Pr[X = k, N = k + l] = Pr[N = k + l] Pr[X = k N = k + l] = by virtue of (2) and (3). Thus, λ k+l (k + l)! e λ ( k + l k ) p k (1 p) l Pr[X = k, Y = l] = λk+l k! l! e λ p k (1 p) l where = λk+l k! l! e λ p k (1 p) l, e λ = e [p+(1 p)]λ = e pλ e (1 p)λ.

Math 151. Rumbos Spring 2014 4 Thus, Pr[X = k, Y = l] = (pλ)k k! e pλ [(1 p)λ]l l! e (1 p)λ = p X (k) p Y (l), in view of (5) and (6). Hence, X and Y are independent. 3. Suppose that the proportion of color blind people in a certain population is 0.005. Estimate the probability that there will be more than one color blind person in a random sample of 600 people from that population. Solution: Set p = 0.005 and n = 600. Denote by Y the number of color blind people in the sample. The, we may assume that Y Binomial(n, p). Since p is small and n is large, we may use the Poisson approximation to the binomial distribution to get where λ = np = 3. Then, Pr[Y = k] λk k! e λ, Pr[Y > 1] = 1 Pr[Y 1] 1 e 3 3e 3 0.800852. Thus, the probability that there will be more than one color blind person in a random sample of 600 people from that population is about 80%. 4. An airline sells 200 tickets for a certain flight on an airplane that has 198 seats because, on average, 1% of purchasers of airline tickets do not appear for departure of their flight. Estimate the probability that everyone who appears for the departure of this flight will have a seat.

Math 151. Rumbos Spring 2014 5 Solution: Set p = 0.01, n = 200 and let Y denote the number of ticket purchasers that do not appear for departure. Then, we may assume that Y Binomial(n, p). We want to estimate the probability Pr[Y 2]. Using the Poisson(λ), with λ = np = 2, approximation to the distribution of Y, we get Pr[Y 2] = 1 Pr[Y 1] 1 e 2 2e 2 0.5940. Thus, the probability that everyone who appears for the departure in this flight will have a seat is about 59.4%. 5. Let X denote a positive random variable such that ln(x) has a Normal(0, 1) distribution. (a) Give the pdf of X and compute its expectation. Solution: Set Z = ln(x), so that Z Normal(0, 1); thus, f Z (y) = 1 2π e z2 /2, for z R. (7) Next, compute the cdf for X, F X (x) = Pr(X x), for x > 0, to get F X (x) = Pr[ln(X) ln(x)] = Pr[Z ln(x)] so that F X (x) = = F Z (ln(x)), { F Z (ln(x)), for x > 0; 0 for x 0. (8) Differentiating (8) with respect to x, for x > 0, we obtain f X (x) = F Z (ln(x)) 1 x,

Math 151. Rumbos Spring 2014 6 so that f X (x) = f Z (ln(x)) 1 x, (9) where we have used the Chain Rule. Combining (7) and (9) yields 1 e (ln x)2 /2, for x > 0; f X (x) = 2π x (10) 0 for x 0. In order to compute the expected value of X, use the pdf in (10) to get E(X) = = 0 xf X (x) dx 1 2π e (ln x)2 /2 dx. (11) Make the change of variables u = ln x in the last integral in (11) to get du = 1 x dx, so that dx = eu du and E(X) = 1 2π e u u2 /2 du. (12) Complete the square in the exponent of the integrand in (12) to obtain E(X) = e 1/2 1 e (u 1)2 /2 du. (13) 2π Next, make the change of variables w = u 1 for the integral in (13) to get (b) Estimate Pr(X 6.5). E(X) = e 1/2 1 2π e w2 /2 dw = e. Solution: Use the result in (8) to compute Pr(X 6.5) = F Z (ln(6.5)), where Z Normal(0, 1). Thus, or about 97%. Pr(X 6.5) = F Z (1.8718) = 0.969383,

Math 151. Rumbos Spring 2014 7 6. Forty seven digits are chosen at random and with replacement from {0, 1, 2,..., 9}. Estimate the probability that their average lies between 4 and 6. Solution: Let X 1, X 2,..., X n, where n = 47, denote the 47 digits. Since the sampling is done without replacement, the random variables X 1, X 2,..., X n are identically uniformly distributed over the digits {0, 1, 2,..., 9} with pmf given by 1, for k = 0, 1, 2,..., 9; 10 p X (k) = (14) 0, elsewhere. Consequently, the mean of the distribution is 9 µ = kp X (k) = 1 9 k = 1 10 10 9 10 = 9 2 2. (15) k=0 k=1 Before we compute the variance, we first compute the second moment of X: 9 9 E(X 2 ) = k 2 p X (k) = k 2 p X (k); k=0 thus, using the pmf of X in (14), E(X 2 ) = 1 9 10 = 1 10 k=1 = 3 (19) 2 = 57 2. Thus, the variance of X is k 2 k=1 9 (9 + 1)(2 9 + 1) 6 σ 2 = E(X 2 ) µ 2 = 57 2 81 4 = 33 4 ;

Math 151. Rumbos Spring 2014 8 so that σ 2 = 8.25. (16) We would like to estimate or where µ is given in (15), so that Pr(4 X n 6), Pr(4 µ X n µ 6 µ), Pr(4 X n 6) = Pr( 0.5 X n µ 1.5) (17) Next, divide the last inequality in (17) by σ/ n, where σ is as given in (16), to get ( Pr(4 X n 6) = Pr 1.19 X n µ σ/ n ) 3.58 (18) For n = 47 can be considered a large sample size, we can apply the Central Limit Theorem to obtain from (18) that Pr(4 X n 6) Pr ( 1.19 Z 3.58), where Z Normal(0, 1). (19) It follows from (19) and the definition of the cdf that Pr(4 X n 6) F Z (3.58) F Z ( 1.19), (20) where F Z is the cdf of Z Normal(0, 1). Using the symmetry of the pdf of Z Normal(0, 1), we can re write (20) as Pr(4 X n 6) F Z (1.19) + F Z (3.58) 1. (21) Finally, using a table of standard normal probabilities, we obtain from (21) that Pr(4 X n 6) 0.8830 + 0.9998 1 = 0.8828. Thus, the probability that the average of the 47 digits is between 4 and 6 is about 88.3%.

Math 151. Rumbos Spring 2014 9 7. Let X 1, X 2,..., X 30 be independent random variables each having a discrete distribution with pmf: 1/4, if x = 0 or x = 2; p(x) = 1/2, if x = 1; 0, otherwise. Estimate the probability that X 1 + X 2 + + X 30 is at most 33. Solution: First, compute the mean, µ = E(X), and variance, σ 2 = Var(X), of the distribution: µ = 0 1 4 + 1 1 2 + 21 4 = 1. (22) σ 2 = E(X 2 ) [E(X)] 2, (23) where E(X 2 ) = 0 2 1 4 + 12 1 2 + 1 22 = 1.5; (24) 4 so that, combining (22), (23) and (24), σ 2 = 1.5 1 = 0.5. (25) Next, let Y = n X k, where n = 30. We would like to estimate k=1 Pr[Y 33], using the continuity correction, Pr[Y 33.5], (26) By the Central Limit Theorem ( ) Y nµ Pr Pr(Z z), n σ for z R, (27) where Z Normal(0, 1), µ = 1, σ 2 = 1.5 and n = 30. It follows from (27) that we can estimate the probability in (26) by Pr[Y 33.5] Pr(Z 0.52) = 0.6985. (28) Thus, according to (28), the probability that X 1 + X 2 + + X 30 is at most 33 is about 70%.

Math 151. Rumbos Spring 2014 10 8. Roll a balanced die 36 times. Let Y denote the sum of the outcomes in each of the 36 rolls. Estimate the probability that 108 Y 144. Suggestion: Since the event of interest is (Y {108, 109,..., 144}), rewrite Pr(108 Y 144) as Pr(107.5 < Y 144.5). Solution: Let X 1, X 2,..., X n, where n = 36, denote the outcomes of the 36 rolls. Since we are assuming that the die is balanced, the random variables X 1, X 2,..., X n are identically uniformly distributed over the digits {1, 2,..., 6}; in other words, X 1, X 2,..., X n is a random sample from the discrete Uniform(6) distribution. Consequently, the mean of the distribution is µ = 6 + 1 = 3.5, (29) 2 and the variance is We also have that σ 2 = (6 + 1)(6 1) 12 Y = n X k, k=1 where n = 36. By the Central Limit Theorem, ( 107.5 nµ Pr(107.5 < Y 144.5) Pr < Z nσ = 35 12. (30) ) 144.5 nµ, nσ (31) where Z Normal(0, 1), n = 36, and µ and σ are given in (29) and (30), respectively. We then have from (31) that Pr(107.5 < Y 144.5) Pr ( 1.81 < Z 1.81) F Z (1.81) F Z ( 1.81) 2F Z (1.81) 1 2(0.9649) 1 0.9298; so that the probability that 108 Y 144 is about 93%.

Math 151. Rumbos Spring 2014 11 9. Let Y Binomial(100, 1/2). Use the Central Limit Theorem to estimate the value of Pr(Y = 50). Solution: We use the continuity correction and estimate Pr(49.5 < Y 50.5). By the Central Limit Theorem, ( 49.5 nµ Pr(49.5 < Y 50.5) Pr < Z nσ where Z Normal(0, 1), n = 100, and nµ = 50 and ( 1 σ = 1 1 ) = 1 2 2 2. We then obtain from (32) that Pr(49.5 < Y 50.5) Pr ( 0.1 < Z 0.1) ) 50.5 nµ, (32) nσ F Z (0.1) F Z ( 0.1) Thus, or about 8%. 2F Z (0.1) 1 2(0.5398) 1 0.0796. Pr(Y = 50) 0.08, 10. Let Y Binomial(n, 0.55). Find the smallest value of n such that, approximately, Pr(Y/n > 1/2) 0.95. (33) Solution: By the Central Limit Theorem, Y n 0.55 (0.55)(1 0.55)/ n D Z Normal(0, 1) as n. (34)

Math 151. Rumbos Spring 2014 12 Thus, according to (33) and (34), we need to find the smallest value of n such that ( ) 0.5 0.55 Pr Z > (0.4975)/ 0.95, n or ( ) n Pr Z > 0.95. (35) 10 The expression in (35) is equivalent to ( ) n 1 Pr Z 0.95, 10 which can be re-written as 1 F Z ( ) n 0.95, (36) 10 where F Z is the cdf of Z Normal(0, 1). By the symmetry of the pdf for Z Normal(0, 1), (36) is equivalent to ( ) n F Z 0.95. (37) 10 The smallest value of n for which (37) holds true occurs when n 10 z, (38) where z is a positive real number with the property F Z (z ) = 0.95. (39) The equality in (39) occurs approximately when z = 1.645. (40) It follows from (38) and (40) that (33) holds approximately when n 1.645, 10 or n 270.6025. Thus, n = 271 is the smallest value of n such that, approximately, Pr(Y/n > 1/2) 0.95.

Math 151. Rumbos Spring 2014 13 11. Let X 1, X 2,..., X n be a random sample from a Poisson distribution with mean n λ. Thus, Y = has a Poisson distribution with mean nλ. Moreover, by i=1 X i the Central Limit Theorem, X = Y/n has, approximately, a Normal(λ, λ/n) distribution for large n. Show that u(y/n) = Y/n is a function of Y/n which is essentially free of λ. Solution: We will show that, for large values of n, the distribution of 2 ( Y n n ) λ (41) is independent of λ. In fact, we will show that, for large values of n, the distribution of the random variables in (41) can be approximated by a Normal(0, 1) distribution. First, note that by the Law of Large Numbers, Y n Pr λ, as n. Thus, for large values of n, we can approximate u(y/n) by its linear approximation around λ ( ) Y u(y/n) u(λ) + u (λ) n λ, (42) where u (λ) = 1 2 λ. (43) Combining (42) and (43) we see that, for large values of n, Y n λ 1 2 n Since, by the Central Limit Theorem, Y n λ. (44) λ n Y n λ λ D Z Normal(0, 1) as n, (45) n

Math 151. Rumbos Spring 2014 14 it follows from (44) and (45) that 2 ( Y n n ) D λ Z Normal(0, 1) as n, which was to be shown. 12. Suppose that factory produces a number X of items in a week, where X can be modeled by a random variable with mean 50. Suppose also that the variance for a week s production is known to be 25. What can be said about the probability that this week s production will be between 40 and 60? Solution: Write where µ = 50, so that Pr(40 < X < 60) = Pr( 10 < X µ < 10), Pr(40 < X < 60) = Pr( X µ < 2σ), where σ = 5. We can then write Pr(40 < X < 60) = 1 Pr( X µ 2σ); where, using Chebyshev s inequality, Consequently, Pr( X µ 2σ) σ2 4σ 2 = 1 4. Pr(40 < X < 60) 1 1 4 = 3 4 ; so that the he probability that this week s production will be between 40 and 60 is at least 75%. 13. Let (X n ) denote a sequence of nonnegative random variables with means µ n = E(X n ), for each n = 1, 2, 3,.... Assume that lim µ n = 0. Show that X n n converges in probability to 0 as n. Solution: We need to show that, for any ε > 0, lim Pr( X n < ε) = 1. (46) n

Math 151. Rumbos Spring 2014 15 We will establish (46) by first applying the Markov inequality: Given a random variable X and any ε > 0, to X = X n, for each n Observe that Pr( X ε) E( X ), (47) ε E( X n ) = E(X n ) = µ n, for all n, since we are assuming that X n is nonnegative. It then follows from (47) that Pr( X n ε) µ n, ε for n = 1, 2, 3,... (48) It follows from (48) that so that Pr( X n < ε) 1 µ n, for n = 1, 2, 3,... ε 1 µ n ε Pr( X n < ε) 1, for n = 1, 2, 3,... (49) Next, use the assumption that lim n µ n = 0 and the Squeeze Lemma to obtain from (49) that lim Pr( X n < ε) = 1, n which is the statement in (46). Hence, X n converges in probability to 0 as n. 14. Let X 1, X 2,..., X n be a random sample from a Poisson distribution with mean n λ. Thus, Y n = has a Poisson distribution with mean nλ. Moreover, by i=1 X i the Central Limit Theorem, X n = Y n /n has, approximately, a Normal(λ, λ/n) distribution for large n. Show that, for large values of n, the distribution of 2 ( Yn n n ) λ is independent of λ. Solution: See solution to Problem 11.