Notes on Continuous Random Variables

Similar documents
Homework 4 Math 11, UCSD, Winter 2018 Due on Tuesday, 13th February

Chapter 4. Continuous Random Variables

HW7 Solutions. f(x) = 0 otherwise. 0 otherwise. The density function looks like this: = 20 if x [10, 90) if x [90, 100]

STAT 516 Midterm Exam 2 Friday, March 7, 2008

MATH 56A: STOCHASTIC PROCESSES CHAPTER 6

Exponential Distribution and Poisson Process

EECS 126 Probability and Random Processes University of California, Berkeley: Fall 2014 Kannan Ramchandran November 13, 2014.

EECS 126 Probability and Random Processes University of California, Berkeley: Fall 2014 Kannan Ramchandran September 23, 2014.

STAT 430/510 Probability Lecture 12: Central Limit Theorem and Exponential Distribution

Random variables. DS GA 1002 Probability and Statistics for Data Science.

3 Continuous Random Variables

Chapter 8: Continuous Probability Distributions

Let X be a continuous random variable, < X < f(x) is the so called probability density function (pdf) if

Midterm Exam 1 Solution

S n = x + X 1 + X X n.

Interlude: Practice Final

Math Fall 2010 Some Old Math 302 Exams There is always a danger when distributing old exams for a class that students will rely on them

(Practice Version) Midterm Exam 2

Guidelines for Solving Probability Problems

Continuous Distributions

2 Continuous Random Variables and their Distributions

Page Max. Possible Points Total 100

CS 5014: Research Methods in Computer Science. Bernoulli Distribution. Binomial Distribution. Poisson Distribution. Clifford A. Shaffer.

STAT/MA 416 Midterm Exam 3 Monday, November 19, Circle the section you are enrolled in:

STA 584 Supplementary Examples (not to be graded) Fall, 2003

Solution: By Markov inequality: P (X > 100) 0.8. By Chebyshev s inequality: P (X > 100) P ( X 80 > 20) 100/20 2 = The second estimate is better.

Lecture 12. Poisson random variables

MATH 407 FINAL EXAM May 6, 2011 Prof. Alexander

2. Suppose (X, Y ) is a pair of random variables uniformly distributed over the triangle with vertices (0, 0), (2, 0), (2, 1).

STAT Chapter 5 Continuous Distributions

STAT/MA 416 Midterm Exam 2 Thursday, October 18, Circle the section you are enrolled in:

ECE 302 Division 2 Exam 2 Solutions, 11/4/2009.

Continuous-time Markov Chains

STAT509: Continuous Random Variable

STAT2201. Analysis of Engineering & Scientific Data. Unit 3

1. If X has density. cx 3 e x ), 0 x < 0, otherwise. Find the value of c that makes f a probability density. f(x) =

1. Let X be a random variable with probability density function. 1 x < f(x) = 0 otherwise

Introduction to Statistical Data Analysis Lecture 3: Probability Distributions

Quick review on Discrete Random Variables

STAT/MA 416 Answers Homework 6 November 15, 2007 Solutions by Mark Daniel Ward PROBLEMS

Northwestern University Department of Electrical Engineering and Computer Science

Since D has an exponential distribution, E[D] = 0.09 years. Since {A(t) : t 0} is a Poisson process with rate λ = 10, 000, A(0.

STAT 430/510 Probability Lecture 16, 17: Compute by Conditioning

Continuous Distributions

Chapter 1: Revie of Calculus and Probability

1 Basic continuous random variable problems

CIVL Continuous Distributions

Midterm Exam 1 (Solutions)

Lecture 4a: Continuous-Time Markov Chain Models

1 Basic continuous random variable problems

CS 1538: Introduction to Simulation Homework 1

Chapter 5. Chapter 5 sections

(b). What is an expression for the exact value of P(X = 4)? 2. (a). Suppose that the moment generating function for X is M (t) = 2et +1 3

Final Exam. Math Su10. by Prof. Michael Cap Khoury

Conditioning a random variable on an event

SDS 321: Introduction to Probability and Statistics

Continuous distributions

CAS Exam MAS-1. Howard Mahler. Stochastic Models

Closed book and notes. 60 minutes. Cover page and four pages of exam. No calculators.

Discrete and Continuous Random Variables

Random Variables. Definition: A random variable (r.v.) X on the probability space (Ω, F, P) is a mapping

Exponential & Gamma Distributions

2905 Queueing Theory and Simulation PART IV: SIMULATION

Probability Distributions Columns (a) through (d)

Brief Review of Probability

Chapter 4 Continuous Random Variables and Probability Distributions

ST 371 (V): Families of Discrete Distributions

Exercises in Probability Theory Paul Jung MA 485/585-1C Fall 2015 based on material of Nikolai Chernov

Chapter 4: Continuous Random Variable

Chapter 8 Queuing Theory Roanna Gee. W = average number of time a customer spends in the system.

Random Variables Example:

Stat410 Probability and Statistics II (F16)

The exponential distribution and the Poisson process

ACCESS TO SCIENCE, ENGINEERING AND AGRICULTURE: MATHEMATICS 2 MATH00040 SEMESTER / Probability

Probability and Statistics Concepts

Discrete Distributions

Chapter 4: Continuous Probability Distributions

Things to remember when learning probability distributions:

Math 180B Homework 9 Solutions

Part 3: Parametric Models

Definition: A random variable X is a real valued function that maps a sample space S into the space of real numbers R. X : S R

Introductory Statistics

Chapter 2. Poisson Processes. Prof. Shun-Ren Yang Department of Computer Science, National Tsing Hua University, Taiwan

December 2010 Mathematics 302 Name Page 2 of 11 pages

Mathematics 375 Probability and Statistics I Final Examination Solutions December 14, 2009

Figure 10.1: Recording when the event E occurs

Part 3: Parametric Models

Continuous random variables

Qualifying Exam CS 661: System Simulation Summer 2013 Prof. Marvin K. Nakayama

Common ontinuous random variables

MATH 3510: PROBABILITY AND STATS July 1, 2011 FINAL EXAM

STOR : Lecture 17. Properties of Expectation - II Limit Theorems

University of Toronto MAT137Y1 Calculus! Test 2 1 December 2017 Time: 110 minutes

Expected Values, Exponential and Gamma Distributions

MA 113 Calculus I Fall 2016 Exam 3 Tuesday, November 15, True/False 1 T F 2 T F 3 T F 4 T F 5 T F. Name: Section:

This does not cover everything on the final. Look at the posted practice problems for other topics.

Continuous Random Variables

CS 237: Probability in Computing

Homework 9 (due November 24, 2009)

Moments. Raw moment: February 25, 2014 Normalized / Standardized moment:

Transcription:

Notes on Continuous Random Variables Continuous random variables are random quantities that are measured on a continuous scale. They can usually take on any value over some interval, which distinguishes them from discrete random variables, which can take on only a sequence of values, usually integers. Typically random variables that represent, for example, time or distance will be continuous rather than discrete. Just as we describe the probability distribution of a discrete random variable by specifying the probability that the random variable takes on each possible value, we describe the probability distribution of a continuous random variable by giving its density function. A density function is a function f which satisfies the following two properties: 1. f(x) for all x. 2. f(x) dx = 1. The first condition says that the density function is always nonnegative, so the graph of the density function always lies on or above the x-axis. The second condition ensures that the area under the density curve is 1. We think of a continuous random variable with density function f as being a random variable that can be obtained by picking a point at random from under the density curve and then reading off the x-coordinate of that point. Because the total area under the density curve is 1, the probability that the random variable takes on a value between a and b is the area under the curve between a and b. More precisely, if X is a random variable with density function f and a < b, then P (a X b) = b a f(x) dx. Example 1: Suppose the income (in tens of thousands of dollars) of people in a community can be approximated by a continuous distribution with density { 2x 2 if x 2 if x < 2 a) Find the probability that a randomly chosen person has an income between $, and $5,. b) Find the probability that a randomly chosen person has an income of at least $,. c) Find the probability that a randomly chosen person has an income of at most $4,. Solution: Let X be the income of a randomly chosen person. The probability that a randomly chosen person has an income between $, and $5, is 5 5 x=5 P ( X 5) = f(x) dx = 2x 2 dx = 2x 1 = 2 ( 5 2 ) = 2 2 5 = 4 15. The probability that a randomly chosen person has an income of at least $, is P (X ) = f(x) dx = = lim n 2x 1 x=n x= = lim n x= n 2x 2 dx = lim 2x 2 dx n ( 2 n + 2 ) = 1. 1

Finally, for part c), we get P (X 4) = 4 f(x) dx = 2 dx + 4 2 x=4 2x 2 dx = 2x 1 = 2 4 + 2 2 = 1 2. Remark 1: Note that part b) required evaluating an improper integral, in which the upper endpoint was infinity. This integral could be evaluated by integrating the density f(x) from to n and taking a limit as n. Evaluation of improper integrals is often required when working with unbounded random variables that can take on any positive number as a value. Remark 2: When the density function of a continuous random variable is defined in two pieces, it is important to be careful about the limits of integration. In part c), we needed to integrate the density from to 4. However, since the density is zero to the left of 2, we only integrated 2x 1 from 2 to 4. It is often useful to draw pictures of the density to avoid mistakenly integrating over the wrong interval. Remark : If we were interested in finding the probability that the random variable X in the Example 1 were exactly equal to, then we would be integrating from to, and we would get zero. This is a general fact about continuous random variables that helps to distinguish them from discrete random variables. A discrete random variable takes on certain values with positive probability. However, if X is a continuous random variable with density f, then P (X = y) = for all y. This may seem counterintuitive at first, since after all X will end up taking some value, but the point is that since X can take on a continuum of values, the probability that it takes on any one particular value is zero. Expected value and standard deviation The procedure for finding expected values and standard deviations for continuous random variables of continuous random variables is similar to the procedure used to calculate expected values and standard deviations for discrete random variables. The differences are that sums in the formula for discrete random variables get replaced by integrals (which are the continuous analogs of sums), while probabilities in the formula for discrete random variables get replaced by densities. More precisely, if X is a random variable with density f(x), then the expected value of X is given by while the variance is given by Note that µ = E[X] = Var(X) = E[(X µ) 2 ] = xf(x) dx, (x µ) 2 f(x) dx. Var(X) = E[(X µ) 2 ] = E[X 2 2µX + µ 2 ] = E[X 2 ] 2µE[X] + µ 2 = E[X 2 ] 2µ 2 + µ 2 = E[X 2 ] µ 2. x=2 2

This formula for the variance (which is valid for all random variables) is often easier to use in computations, so we may calculate ( ) Var(X) = E[X 2 ] µ 2 = x 2 f(x) dx µ 2. As in the case of discrete random variables, the standard deviation of X is the square root of the variance of X. Example 2: Suppose a train arrives shortly after 1: PM each day, and that the number of minutes after 1: that the train arrives can be modeled as a continuous random variable with density { 2(1 x) if x 1 Find the mean and standard deviation of the number of minutes after 1: that the train arrives. Solution: Let X be the number of minutes after 1: that the train arrives. The mean (or, equivalently, the expected value) of X is given by µ = E[X] = Also, we have E[X 2 ] = Therefore, xf(x) dx = x 2 f(x) dx = x 2(1 x) dx = x 2 2(1 x) = Var(X) = 1 ( 1 and the standard deviation is 1/18.24. 2x 2x 2 dx = ) (x 2 2x x=1 = 1 x=. ( ) 2x 2x 2 2x dx = 2x4 x=1 = 2 4 2 4 = 1. ) 2 = 1 1 9 = 1 18, Special Continuous Distributions As was the case with discrete random variables, when we gave special attention to the geometric, binomial, and Poisson distributions, some continuous distributions occur repeatedly in applications. Probably the three most important continuous distributions are the uniform distribution, the exponential distribution, and the normal distribution. Uniform Distribution: If a < b, then we say a random variable X has the uniform distribution on [a, b] if { 1 b a if a x b Note that the density function is zero outside [a, b], so a random variable having the uniform distribution on [a, b] always falls between a and b. Because the density is flat between a and b, we can think of the uniform distribution as representing a number chosen uniformly at random from the interval [a, b]. x=

Example : If X has the uniform distribution on [2, 5], calculate P (X > 4). Solution: The density of X is given by Therefore, P (X 4) = 4 f(x) dx = { 1 if 2 x 5 5 4 1 dx = x x=5 x=4 = 5 4 = 1. Alternatively, we could observe that the area under the density function between x = 4 and x = 5 is a rectangle of length 1 and height 1/, which has area 1/. Exponential Distribution: If λ >, then we say a random variable X has the exponential distribution with rate λ if its density is given by { λe λx if x The exponential distribution is a model of the amount of time we have to wait for an event that happens at rate λ per unit time. The exponential distribution should also describe the time we have to wait for a customer to arrive in a store, a goal to be scored in a hockey game, a radioactive particle to be emitted, an accident to occur on a roadway, or an earthquake to occur in California. The exponential distribution arises in the same circumstances under which the Poisson distribution arises. However, whereas the number of events that occur in a certain time interval has the Poisson distribution, the amount of time to wait for the next event has the exponential distribution. Because the Poisson distribution is integer-valued and the exponential distribution is a continuous distribution, one should be able to keep the two distributions straight. The most important property of the exponential distribution is known as the memoryless property, which says that if the time to wait for an event to occur has the exponential distribution, then the probability that we have to wait an additional time t is the same no matter how long we have already waited. More formally, if X has an exponential distribution with rate λ, then P (X s + t X s) = P (X t). That is, the probability that we have to wait for an additional time t (and therefore a total time of s + t) given that we have already waited for time s is the same as the probability at the start that we would have had to wait for time t. This is true for all s, that is, no matter how long we have already waited. To see why the memoryless property holds, note that for all t, we have It follows that P (X t) = P (X s + t X s) = t λe λx dx = e λx = e λt. P (X s + t and X s) P (X s) = e λ(s+t) e λs = e λs e λt e λs 4 t = P (X s + t) P (X s) = e λt = P (X t),

which verifies the memoryless property. If X has an exponential distribution with parameter λ, then E[X] = 1/λ. Thus, the expected time to wait for an event to occur is inversely proportional to the rate at which the event occurs, as would be expected. Example 4: If customers arrive in a store at the rate of 4 per hour, what is the probability that we will have to wait between 15 and minutes for the next customer to arrive. Solution: Let X be the time we have to wait for the next customer to arrive. Then X has the exponential distribution with parameter λ = 4, so X has density { 4e 4x if x The probability that we have to wait between 15 and minutes is P (1/4 X 1/2) = /2 1/4 x=1/2 4e 4x dx = e 4x = e 4/2 ( e 4/4 ) = e 1 e 2.2. x=1/4 Normal Distribution: A random variable has a normal distribution with mean µ and standard deviation σ if its density is given by 1 σ /2σ2 e (x µ)2. 2π The normal distribution is the most important distribution in statistics, arising in numerous applications because of a famous result known as the Central Limit Theorem. Because the normal distribution is discussed in our textbook, we do not pursue it further here. Problems 1: Suppose X is a random variable with density { 2x if < x < 1. a) Find P (X 1/2). b) Find P (X /4). c) Find P (X 2). d) Find E[X]. e) Find the standard deviation of X. 2: Suppose X is a random variable with density { cx 2 if < x < 2 for some positive number c. What is c? 5

: Suppose the number of years that a computer lasts has density { 8x if x 2. a) Find the probability that the computer lasts between and 5 years. b) Find the probability that the computer lasts at least 4 years. c) Find the probability that the computer lasts less than 1 year. d) Find the probability that the computer lasts exactly 2.48 years. e) Find the expected value of the number of years that the computer lasts. 4: Suppose that the number of hours that it takes for a student to finish an exam has density { 2 5 (x + 1) if 1 < x < 2. a) Find the probability that the student finishes the exam in less than 1.5 hours. b) Find the mean and standard deviation of the number of hours it takes to finish the exam. 5: Suppose that if you arrive at a bus stop at 8:, the number of minutes that you will have to wait for the next bus is uniformly distributed on [, 1]. a) Find the probability that you will have to wait at least six minutes for the bus. b) Find the probability that you will have to wait between two and four minutes. c) Find the expected value and standard deviation of the number of minutes that you have to wait for the next bus. : Suppose the random variable X has the uniform distribution on [a, b]. Find expressions involving a and b for the expected value, variance, and standard deviation of X. Check that your expressions when a = and b = 1 agree with what you got in part c) of problem 5. 7: Suppose major earthquakes in a certain region occur independently of one another at the rate of one every ten years. Find the probability that the next major earthquake will occur between 7 and 12 years from now. 8: Suppose that the amount of time (in months) that a light bulb lasts before it burns out has an exponential distribution with parameter λ = 1/5. a) What is the probability that it lasts at least three months? b) If it has already lasted two months, what is the probability that it will last at least three more months? c) On average, how many months will the light bulb last? 9: Accidents occur at a busy intersection at the rate of two per year. What is the probability that it will be at least one year before the next accident at the intersection? Compute the answer using the following two methods: a) Let X be the number of accidents in the next year. Find the distribution of X and calculate P (X = ). b) Let T be the amount of time until the next accident. Find the distribution of T and calculate P (T > 1).