Distributions of Functions of Random Variables

Similar documents
HW4 : Bivariate Distributions (1) Solutions

STAT/MATH 395 A - PROBABILITY II UW Winter Quarter Moment functions. x r p X (x) (1) E[X r ] = x r f X (x) dx (2) (x E[X]) r p X (x) (3)

STAT/MATH 395 PROBABILITY II

Bivariate Distributions

2 Functions of random variables

STAT/MATH 395 PROBABILITY II

Distributions of Functions of Random Variables. 5.1 Functions of One Random Variable

Math 3215 Intro. Probability & Statistics Summer 14. Homework 5: Due 7/3/14

Probability and Distributions

Continuous Random Variables

Problem Y is an exponential random variable with parameter λ = 0.2. Given the event A = {Y < 2},

1 Joint and marginal distributions

1 Review of Probability and Distributions

Lecture 3: Random variables, distributions, and transformations

2. The CDF Technique. 1. Introduction. f X ( ).

2 (Statistics) Random variables

Chapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued

Review Quiz. 1. Prove that in a one-dimensional canonical exponential family, the complete and sufficient statistic achieves the

HW3 : Moment functions Solutions

Mathematical Statistics 1 Math A 6330

Check Your Understanding of the Lecture Material Finger Exercises with Solutions

Discrete Random Variables (1) Solutions

Lecture 1: August 28

MITOCW MITRES6_012S18_L23-05_300k

MTH739U/P: Topics in Scientific Computing Autumn 2016 Week 6

Multiple Random Variables

Chapter 5,6 Multiple RandomVariables

Random Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R

Transformations and Expectations

Chapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued

p. 6-1 Continuous Random Variables p. 6-2

County Council Named for Kent

Quick review on Discrete Random Variables

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

ECE 302 Division 2 Exam 2 Solutions, 11/4/2009.

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

MATH/STAT 395 Winter 2013 This homework is due at the beginning of class on Friday, March 1.

Lecture 4. Continuous Random Variables and Transformations of Random Variables

Solution to Assignment 3

Probability Models. 4. What is the definition of the expectation of a discrete random variable?

4. CONTINUOUS RANDOM VARIABLES

Distribution of Random Samples & Limit theorems

Lecture 11. Probability Theory: an Overveiw

Chapter 4 Multiple Random Variables

CS145: Probability & Computing Lecture 11: Derived Distributions, Functions of Random Variables

Will Landau. Feb 21, 2013

3 Operations on One Random Variable - Expectation

CS145: Probability & Computing

' '-'in.-i 1 'iritt in \ rrivfi pr' 1 p. ru

Transformation of Probability Densities

Probability Distribution And Density For Functional Random Variables

Ch3. Generating Random Variates with Non-Uniform Distributions

STA2603/205/1/2014 /2014. ry II. Tutorial letter 205/1/

4 Pairs of Random Variables

Properties of Continuous Probability Distributions The graph of a continuous probability distribution is a curve. Probability is represented by area

Continuous random variables

Measure-theoretic probability

ENSC327 Communications Systems 19: Random Processes. Jie Liang School of Engineering Science Simon Fraser University

When Are Two Random Variables Independent?

Probability (continued)

Random Variables. Saravanan Vijayakumaran Department of Electrical Engineering Indian Institute of Technology Bombay

Math 416 Lecture 2 DEFINITION. Here are the multivariate versions: X, Y, Z iff P(X = x, Y = y, Z =z) = p(x, y, z) of X, Y, Z iff for all sets A, B, C,

Bivariate distributions

Multivariate random variables

STAT 430/510: Lecture 15

MAS223 Statistical Inference and Modelling Exercises

Chapter 4. Continuous Random Variables 4.1 PDF

1 Expectation of a continuously distributed random variable

Review: mostly probability and some statistics

Basics of Stochastic Modeling: Part II

E X A M. Probability Theory and Stochastic Processes Date: December 13, 2016 Duration: 4 hours. Number of pages incl.

Probability and Statistics Notes

Lecture Notes 3 Multiple Random Variables. Joint, Marginal, and Conditional pmfs. Bayes Rule and Independence for pmfs

STAT 801: Mathematical Statistics. Distribution Theory

Probability Review. Yutian Li. January 18, Stanford University. Yutian Li (Stanford University) Probability Review January 18, / 27

Introduction to Statistical Inference Self-study

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.

Continuous Random Variables

Continuous Random Variables

Measure theory and stochastic processes TA Session Problems No. 3

Math 416 Lecture 3. The average or mean or expected value of x 1, x 2, x 3,..., x n is

UCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, Practice Final Examination (Winter 2017)

Random Signal Transformations and Quantization

STAT 450: Statistical Theory. Distribution Theory. Reading in Casella and Berger: Ch 2 Sec 1, Ch 4 Sec 1, Ch 4 Sec 6.

Generating Random Variables

Introduction to Probability Theory

Estimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators

Closed book and notes. 60 minutes. Cover page and four pages of exam. No calculators.

Recitation 2: Probability

4. Distributions of Functions of Random Variables

Chapter 2: Random Variables

Math 151. Rumbos Spring Solutions to Review Problems for Exam 2

1 Random Variable: Topics

Multivariate Distribution Models

SOLUTION FOR HOMEWORK 12, STAT 4351

The random variable 1

Master s Written Examination - Solution

1.1 Review of Probability Theory

General Random Variables

RFID & Item-Level Information Visibility

Transcription:

STAT/MATH 395 A - PROBABILITY II UW Winter Quarter 217 Néhémy Lim Distributions of Functions of Random Variables 1 Functions of One Random Variable In some situations, you are given the pdf f X of some rrv X. But you may actually be interested in some function of the initial rrv : Y = u(x). In this chapter, we are going to study different techniques for finding the distribution of functions of random variables. 1.1 Distribution Function Technique Assume that we are given a continuous rrv X with pdf f X. We want to find the pdf of Y = u(x). As seen previously when we studied the exponential distribution, we can apply the following strategy : 1. First, find the cdf (cumulative distribution function) F Y (y) 2. Then, differentiate the cumulative distribution function F Y (y) to get the probability density function f Y (y). That is: f Y (y) = F Y (y) Example 1. Let X be a rrv with pdf : f X (x) = 3x 2 1 (,1) (x) What is the pdf of Y = X 2? Answer. The cdf of Y is : for y (, 1) = P(X 2 y) Note that the transformation u : x x 2 is strictly increasing on (, 1). Thus, u is invertible and its inverse v : y y is also strictly increasing. Therefore, for y (, 1), we have F Y (y) = P(v(X 2 ) v(y)) = P(X y) = F X ( y) = y = [ x 3] y = y 3/2 3x 2 dx 1

Hence, the pdf of Y is obtained as follows: for y (, 1), = 3 2 y3/2 1 In a nutshell, f Y (y) = 3 2 y1/2 1 (,1) (y) Example 2. Let X be a rrv with pdf : f X (x) = 3(1 x) 2 1 (,1) (x) What is the pdf of Y = (1 X) 3? Answer. The cdf of Y is : for y (, 1) = P((1 X) 3 y) Note that the transformation u : x (1 x) 3 is strictly decreasing on (, 1). Thus, u is invertible and its inverse v : y 1 y 1/3 is also strictly decreasing. Therefore, for y (, 1), we have F Y (y) = P(v((1 X) 3 ) v(y)) = P(X 1 y 1/3 ) = 1 F X (1 y 1/3 ) = 1 1 y 1/3 3(1 x) 2 dx = 1 [ (1 x) 3] 1 y 1/3 ( ( ) 3 = 1 + 1 (1 y 1/3 ) (1 ) 3) = y Hence, the pdf of Y is obtained as follows: for y (, 1), = 1 In a nutshell, f Y (y) = 1 (,1) (y) That is Y follows a uniform distribution on (, 1). 2

1.2 Change-of-Variable Technique Theorem 1.1. Let X be a continuous random variable on probability space (Ω, A, P) with pdf f X = f 1 S where S is the support of f X. If u is strictly monotonic with inverse function v, then the pdf of random variable Y = u(x) is given by : f Y (y) = f (v(y)) v (y) 1 u(s) (y) (1) Proof. Assume u is strictly increasing. Then, u is invertible and its inverse v is also strictly increasing. Therefore, = P(u(X) y) = P(X v(y)) = F X (v(y)) = d dy F X(v(y)) = F X(v(y))v (y) = f X (v(y))v (y) On the other hand, assume u is strictly decreasing. Then, u is invertible and its inverse v is also strictly decreasing. Therefore, = P(u(X) y) = P(X v(y)) = 1 F X (v(y)) = d dy {1 F X(v(y))} = F X(v(y))v (y) = f X (v(y))v (y) We can merge the two cases since v (y) if v is increasing and v (y) if v is decreasing. 3

For illustration, apply the Change-of-Variable Technique to Examples 1 and 2 and make sure you find the same results. Case of two-to-one transformations. Example 3. Let X be a rrv with pdf : f X (x) = x2 3 1 ( 1,2)(x) What is the pdf of Y = X 2? Answer. Note that the transformation u : x x 2 is not strictly monotonic on ( 1, 2). Therefore we cannot apply Theorem 1.1 straight away. More precisely, u is two-to-one on ( 1, 1) and one-to-one on (1, 2). Let us focus on the interval ( 1, 1) and use the distribution technique. In that case, we have for y (, 1) : Noting that = P(X 2 y) = P( y X y) = F X ( y) F X ( y) u is strictly decreasing on ( 1, ) with strictly decreasing inverse v 1 : y y u is strictly increasing on (, 1) with strictly increasing inverse v 2 : y y and by differentiating the cdf, we obtain : for y (, 1), = v 2(y)f X (v 2 (y)) + ( v 1(y))f X (v 1 (y)) = 1 2 y 1/2 f X ( y) + 1 2 y 1/2 f X ( y) = 1 y 2 2 y 1/2 3 + 1 ( y) 2 y 1/2 2 3 y = 3 On the interval (1, 2), u is strictly increasing, thus we can apply Theorem 1.1. After some calculations, you should find that for y (1, 4), y f Y (y) = 6 In a nutshell, y/3 if < y < 1 f Y (y) = y/6 if 1 < y < 4 otherwise 4

Let us generalize our finding. If the transformation u is two-to-one on some interval and can be split into two strictly monotonic functions with inverses v 1 and v 2. The the pdf of Y = u(x) on that interval is : f Y (y) = v 1(y) f X (v 1 (y)) + v 2(y) f X (v 2 (y)) 2 Transformations of Two Random Variables Theorem 2.1. Let X and Y be two continuous random variables on probability space (Ω, A, P) with joint pdf f XY = f 1 S where S R 2 is the support of f XY. If u = (u 1, u 2 ) is an invertible function on S with inverse function v = (v 1, v 2 ), then the joint pdf of random variables W = u 1 (X, Y ) and Z = u 2 (X, Y ) is given by : f W Z (w, z) = f (v 1 (w, z), v 2 (w, z)) J 1 u(s) (w, z) (2) where J is the Jacobian of v at point (s, t) defined by the following determinant : v 1(w,z) v 1(w,z) J = w z = v 1(w, z) v 2 (w, z) v 2(w, z) v 1 (w, z) w z w z v 2(w,z) w v 2(w,z) z Example 4. Let X and Y be 2 rrv with joint pdf : f XY (x, y) = e (x+y) 1 (, ) 2(x, y) X X+Y? What is the joint pdf of W = X + Y and Z = Answer. Let us solve the following system for X and Y : { W = u1 (X, Y ) = X + Y Z = u 2 (X, Y ) = X X+Y { Y = W X Z = X W { Y = W X X = W Z { Y = W W Z = v2 (W, Z) X = W Z = v 1 (W, Z) The determinant of the Jacobian of v = (v 1, v 2 ) is thus given by : v 1(w,z) v 1(w,z) J = w z v 2(w,z) v 2(w,z) w z = z w 1 z w = wz w(1 z) = w 5

The support of the joint pdf of X and Y is S = (, ) 2. The transformation u : (x, y) (x + y, x/(x + y)) maps S in the xy-plane into the domain u(s) in the (w, z)-plane given by w = x + y > and z = x/(x + y) (, 1). Thus, the joint pdf of W and Z is given by : f W Z (w, z) = e (v1(w,z)+v2(w,z)) w 1 (, ) (,1) (w, z) = e (wz+w wz) w1 (, ) (,1) (w, z) = we w 1 (, ) (,1) (w, z) 6