Probability Theory and Statistics (EE/TE 3341) Homework 3 Solutions

Similar documents
Lecture 4: Bernoulli Process

BINOMIAL DISTRIBUTION

Continuous Random Variables

Probability and Stochastic Processes: A Friendly Introduction for Electrical and Computer Engineers Roy D. Yates and David J.

Continuous-Valued Probability Review

Notes for Math 324, Part 17

Math 218 Supplemental Instruction Spring 2008 Final Review Part A

Problems for 2.6 and 2.7

3.4. The Binomial Probability Distribution

Random Variables Example:

Relationship between probability set function and random variable - 2 -

EECS 126 Probability and Random Processes University of California, Berkeley: Fall 2014 Kannan Ramchandran November 13, 2014.

Introduction to Probability 2017/18 Supplementary Problems

Chapter 2 Random Variables

DISCRETE VARIABLE PROBLEMS ONLY

CHAPTER 6. 1, if n =1, 2p(1 p), if n =2, n (1 p) n 1 n p + p n 1 (1 p), if n =3, 4, 5,... var(d) = 4var(R) =4np(1 p).

A Probability Primer. A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes.

Sample Space: Specify all possible outcomes from an experiment. Event: Specify a particular outcome or combination of outcomes.

Discrete Random Variable

Topic 3 - Discrete distributions

Exam 3, Math Fall 2016 October 19, 2016

Page Max. Possible Points Total 100

Guidelines for Solving Probability Problems

Math Treibergs. Solutions to Fourth Homework February 13, 2009

Discrete Random Variables (cont.) Discrete Distributions the Geometric pmf

Statistics for Economists. Lectures 3 & 4

CMPSCI 240: Reasoning Under Uncertainty

Notes on Continuous Random Variables

Random variable X is a mapping that maps each outcome s in the sample space to a unique real number x, x. X s. Real Line

Interlude: Practice Final

How are Geometric and Poisson probability distributions different from the binomial probability distribution? How are they the same?

IEOR 3106: Introduction to Operations Research: Stochastic Models. Professor Whitt. SOLUTIONS to Homework Assignment 1

More Discrete Distribu-ons. Keegan Korthauer Department of Sta-s-cs UW Madison

Class 26: review for final exam 18.05, Spring 2014

Lecture 3. Discrete Random Variables

CS 1538: Introduction to Simulation Homework 1

Geometric Distribution The characteristics of a geometric experiment are: 1. There are one or more Bernoulli trials with all failures except the last

1. A machine produces packets of sugar. The weights in grams of thirty packets chosen at random are shown below.

ISyE 6739 Test 1 Solutions Summer 2015

Homework 4 Math 11, UCSD, Winter 2018 Due on Tuesday, 13th February

Massachusetts Institute of Technology

Introductory Probability

Entropy. Probability and Computing. Presentation 22. Probability and Computing Presentation 22 Entropy 1/39

Fault-Tolerant Computer System Design ECE 60872/CS 590. Topic 2: Discrete Distributions

Discrete Random Variable Practice

Descriptive Statistics and Probability Test Review Test on May 4/5

HW7 Solutions. f(x) = 0 otherwise. 0 otherwise. The density function looks like this: = 20 if x [10, 90) if x [90, 100]

ELEG 3143 Probability & Stochastic Process Ch. 2 Discrete Random Variables

Lecture 8 : The Geometric Distribution

Part 3: Parametric Models

Introduction to Probability

Common Discrete Distributions

STATISTICAL THINKING IN PYTHON I. Probabilistic logic and statistical inference

Random Variables. Definition: A random variable (r.v.) X on the probability space (Ω, F, P) is a mapping

ASM Study Manual for Exam P, Second Edition By Dr. Krzysztof M. Ostaszewski, FSA, CFA, MAAA Errata

Errata for the ASM Study Manual for Exam P, Fourth Edition By Dr. Krzysztof M. Ostaszewski, FSA, CFA, MAAA

Answers to selected exercises

Introduction to Probability, Fall 2013

(b). What is an expression for the exact value of P(X = 4)? 2. (a). Suppose that the moment generating function for X is M (t) = 2et +1 3

MATH : EXAM 2 INFO/LOGISTICS/ADVICE

IE 336 Seat # Name (clearly) Closed book. One page of hand-written notes, front and back. No calculator. 60 minutes.

1 Inverse Transform Method and some alternative algorithms

CS 237 Fall 2018, Homework 06 Solution

Lecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes

Chapters 3.2 Discrete distributions

Expectations. Definition Let X be a discrete rv with set of possible values D and pmf p(x). The expected value or mean value of X, denoted by E(X ) or

Chapter Three: Translations & Word Problems

n(1 p i ) n 1 p i = 1 3 i=1 E(X i p = p i )P(p = p i ) = 1 3 p i = n 3 (p 1 + p 2 + p 3 ). p i i=1 P(X i = 1 p = p i )P(p = p i ) = p1+p2+p3

8th Grade Standards for Algebra Readiness Solve one-variable linear equations. 3. d 8 = 6 4. = 16 5.

ECE353: Probability and Random Processes. Lecture 5 - Cumulative Distribution Function and Expectation

MAS275 Probability Modelling Exercises

1. If X has density. cx 3 e x ), 0 x < 0, otherwise. Find the value of c that makes f a probability density. f(x) =

ECE 302 Division 2 Exam 2 Solutions, 11/4/2009.

To find the median, find the 40 th quartile and the 70 th quartile (which are easily found at y=1 and y=2, respectively). Then we interpolate:

SS257a Midterm Exam Monday Oct 27 th 2008, 6:30-9:30 PM Talbot College 342 and 343. You may use simple, non-programmable scientific calculators.

ECE 313: Conflict Final Exam Tuesday, May 13, 2014, 7:00 p.m. 10:00 p.m. Room 241 Everitt Lab

ASM Study Manual for Exam P, First Edition By Dr. Krzysztof M. Ostaszewski, FSA, CFA, MAAA Errata

Bernoulli and Binomial

CHAPTER 2. Solution to Problem 2.1. the weekend. We have. Let X be the number of points the MIT team earns over

Discrete Distributions

Find the value of n in order for the player to get an expected return of 9 counters per roll.

6.1 Quadratic Expressions, Rectangles, and Squares. 1. What does the word quadratic refer to? 2. What is the general quadratic expression?

What is Probability? Probability. Sample Spaces and Events. Simple Event

ECE Homework Set 2

The story of the film so far... Mathematics for Informatics 4a. Continuous-time Markov processes. Counting processes

Taylor and Maclaurin Series. Approximating functions using Polynomials.

Probability Theory and Simulation Methods. April 6th, Lecture 19: Special distributions

Stats for Engineers: Lecture 4

Example 1. The sample space of an experiment where we flip a pair of coins is denoted by:

Some Practice Questions for Test 2

STEP Support Programme. Statistics STEP Questions: Solutions

REPEATED TRIALS. p(e 1 ) p(e 2 )... p(e k )

Conditional Probability

Chapter 2: Discrete Distributions. 2.1 Random Variables of the Discrete Type

STAT2201. Analysis of Engineering & Scientific Data. Unit 3

Basic concepts of probability theory

Learning Objectives for Stat 225

STAT/MATH 395 PROBABILITY II

Random Variable. Discrete Random Variable. Continuous Random Variable. Discrete Random Variable. Discrete Probability Distribution

1 Gambler s Ruin Problem

Transcription:

Probability Theory and Statistics (EE/TE 3341) Homework 3 Solutions Yates and Goodman 3e Solution Set: 3.2.1, 3.2.3, 3.2.10, 3.2.11, 3.3.1, 3.3.3, 3.3.10, 3.3.18, 3.4.3, and 3.4.4 Problem 3.2.1 Solution (a) We wish to find the value of c that makes the PMF sum up to one. c(1/2) n n = 0, 1, 2, P N (n) = Therefore, 2 n=0 P N(n) = c + c/2 + c/4 = 1, implying c = 4/7. (b) The probability that N 1 is P [N 1] = P [N = 0] + P [N = 1] = 4/7 + 2/7 = 6/7. (2) Problem 3.2.3 Solution (a) We choose c so that the PMF sums to one. x P X (x) = c 2 + c 4 + c 8 = 7c 8 = 1. Thus c = 8/7. (b) P [X = 4] = P X (4) = 8 7 4 = 2 7. (2) 1

(c) P [X < 4] = P X (2) = 8 7 2 = 4 7. (3) (d) Problem 3.2.10 Solution P [3 X 9] = P X (4) + P X (8) = 8 7 4 + 8 7 8 = 3 7. (4) From the problem statement, a single is twice as likely as a double, which is twice as likely as a triple, which is twice as likely as a home-run. If p is the probability of a home run, then P B (4) = p P B (3) = 2p P B (2) = 4p P B = 8p Since a hit of any kind occurs with probability of.300, p + 2p + 4p + 8p = 0.300 which implies p = 0.02. Hence, the PMF of B is 0.70 b = 0, 0.16 b = 1, 0.08 b = 2, P B (b) = 0.04 b = 3, 0.02 b = 4, (2) Problem 3.2.11 Solution (a) In the setup of a mobile call, the phone will send the SETUP message up to six times. Each time the setup message is sent, we have a Bernoulli trial with success probability p. Of course, the phone stops trying as soon as there is a success. Using r to denote a successful response, and n a non-response, the sample tree is 2

p r K=1 p r K=2 p r K=3 p r K=4 p r K=5 p r K=6 1 p n K=6 (b) We can write the PMF of K, the number of SETUP messages sent as (1 p) k 1 p k = 1, 2,..., 5, P K (k) = (1 p) 5 p + (1 p) 6 = (1 p) 5 k = 6, Note that the expression for P K (6) is different because K = 6 if either there was a success or a failure on the sixth attempt. In fact, K = 6 whenever there were failures on the first five attempts which is why P K (6) simplifies to (1 p) 5. (c) Let B denote the event that a busy signal is given after six failed setup attempts. The probability of six consecutive failures is P[B] = (1 p) 6. (d) To be sure that P[B] 0.02, we need p 1 (0.02) 1/6 = 0.479. Problem 3.3.1 Solution (a) If it is indeed true that Y, the number of yellow M&M s in a package, is uniformly distributed between 5 and 15, then the PMF of Y, is 1/11 y = 5, 6, 7,..., 15 P Y (y) = 0 otherwise (b) (c) (d) P [Y < 10] = P Y (5) + P Y (6) + + P Y (9) = 5/11. (2) P [Y > 12] = P Y (13) + P Y (14) + P Y (15) = 3/11. (3) P [8 Y 12] = P Y (8) + P Y (9) + + P Y (12) = 5/11. (4) 3

Problem 3.3.3 Solution (a) Each paging attempt is an independent Bernoulli trial with success probability p. The number of times K that the pager receives a message is the number of successes in n Bernoulli trials and has the binomial PMF ( n ) P K (k) = k p k (1 p) n k k = 0, 1,..., n, (b) Let R denote the event that the paging message was received at least once. The event R has probability P [R] = P [B > 0] = 1 P [B = 0] = 1 (1 p) n. (2) To ensure that P[R] 0.95 requires that n ln(0.05)/ ln(1 p). For p = 0.8, we must have n 1.86. Thus, n = 2 pages would be necessary. Problem 3.3.10 Solution Since an average of T/5 buses arrive in an interval of T minutes, buses arrive at the bus stop at a rate of 1/5 buses per minute. (a) From the definition of the Poisson PMF, the PMF of B, the number of buses in T minutes, is (T/5) b e T/5 /b! b = 0, 1,..., P B (b) = (b) Choosing T = 2 minutes, the probability that three buses arrive in a two minute interval is P B (3) = (2/5) 3 e 2/5 /3! 0.0072. (2) (c) By choosing T = 10 minutes, the probability of zero buses arriving in a ten minute interval is P B (0) = e 10/5 /0! = e 2 0.135. (3) 4

(d) The probability that at least one bus arrives in T minutes is P [B 1] = 1 P [B = 0] = 1 e T/5 0.99. (4) Rearranging yields T 5 ln 100 23.0 minutes. Problem 3.3.18 Solution (a) Let S n denote the event that the Sixers win the series in n games. Similarly, C n is the event that the Celtics in in n games. The Sixers win the series in 3 games if they win three straight, which occurs with probability P [S 3 ] = (1/2) 3 = 1/8. The Sixers win the series in 4 games if they win two out of the first three games and they win the fourth game so that ( ) 3 P [S 4 ] = (1/2) 3 (1/2) = 3/16. (2) 2 The Sixers win the series in five games if they win two out of the first four games and then win game five. Hence, ( ) 4 P [S 5 ] = (1/2) 4 (1/2) = 3/16. (3) 2 By symmetry, P[C n ] = P[S n ]. Further we observe that the series last n games if either the Sixers or the Celtics win the series in n games. Thus, P [N = n] = P [S n ] + P [C n ] = 2 P [S n ]. (4) Consequently, the total number of games, N, played in a best of 5 series between the Celtics and the Sixers can be described by the PMF 2(1/2) 3 = 1/4 n = 3, 2 ( 3 P N (n) = 1) (1/2) 4 = 3/8 n = 4, 2 ( 4 2) (1/2) 5 (5) = 3/8 n = 5, 5

(b) For the total number of Celtic wins W, we note that if the Celtics get w < 3 wins, then the Sixers won the series in 3 + w games. Also, the Celtics win 3 games if they win the series in 3,4, or 5 games. Mathematically, P [S 3+w ] w = 0, 1, 2, P [W = w] = (6) P [C 3 ] + P [C 4 ] + P [C 5 ] w = 3. Thus, the number of wins by the Celtics, W, has the PMF shown below. P [S 3 ] = 1/8 w = 0, P [S 4 ] = 3/16 w = 1, P W (w) = P [S 5 ] = 3/16 w = 2, 1/8 + 3/16 + 3/16 = 1/2 w = 3, (7) (c) The number of Celtic losses L equals the number of Sixers wins W S. This implies P L (l) = P WS (l). Since either team is equally likely to win any game, by symmetry, P WS (w) = P W (w). This implies P L (l) = P WS (l) = P W (l). The complete expression of for the PMF of L is 1/8 l = 0, Problem 3.4.3 Solution 3/16 l = 1, P L (l) = P W (l) = 3/16 l = 2, 1/2 l = 3, (8) (a) Similar to the previous problem, the graph of the CDF is shown below. 0 x < 3, 0.4 3 x < 5, F X (x) = 0.8 5 x < 7, 1 x 7. F X (x) 0.8 1 0.6 0.4 0.2 0 3 0 5 7 x 6

(b) The corresponding PMF of X is 0.4 x = 3 0.4 x = 5 P X (x) = 0.2 x = 7 0 otherwise (2) Problem 3.4.4 Solution Let q = 1 p, so the PMF of the geometric (p) random variable K is pq k 1 k = 1, 2,..., P K (k) = For any integer k 1, the CDF obeys F K (k) = k P K (j) = j=1 k pq j 1 = 1 q k. (2) Since K is integer valued, F K (k) = F K ( k ) for all integer and non-integer values of k. (If this point is not clear, you should review Example 3.22.) Thus, the complete expression for the CDF of K is 0 k < 1, F K (k) = 1 (1 p) k (3) k 1. j=1 7