Random Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R

Similar documents
Random Variables. Saravanan Vijayakumaran Department of Electrical Engineering Indian Institute of Technology Bombay

Random Variables. Cumulative Distribution Function (CDF) Amappingthattransformstheeventstotherealline.

Continuous Random Variables

Chapter 2. Some Basic Probability Concepts. 2.1 Experiments, Outcomes and Random Variables

Probability and Distributions

2 (Statistics) Random variables

P (x). all other X j =x j. If X is a continuous random vector (see p.172), then the marginal distributions of X i are: f(x)dx 1 dx n

Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016

Chapter 4. Chapter 4 sections

BASICS OF PROBABILITY

3. Probability and Statistics

Joint Distribution of Two or More Random Variables

p. 6-1 Continuous Random Variables p. 6-2

More on Distribution Function

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems

1 Review of Probability and Distributions

ECON Fundamentals of Probability

18.440: Lecture 28 Lectures Review

Joint Distributions. (a) Scalar multiplication: k = c d. (b) Product of two matrices: c d. (c) The transpose of a matrix:

1 Presessional Probability

18.440: Lecture 28 Lectures Review

MAS113 Introduction to Probability and Statistics. Proofs of theorems

Arkansas Tech University MATH 3513: Applied Statistics I Dr. Marcel B. Finan

Algorithms for Uncertainty Quantification

1 Random Variable: Topics

5 Operations on Multiple Random Variables

Lecture 2: Repetition of probability theory and statistics

1.1 Review of Probability Theory

Probability- the good parts version. I. Random variables and their distributions; continuous random variables.

Review of Probability Theory

PCMI Introduction to Random Matrix Theory Handout # REVIEW OF PROBABILITY THEORY. Chapter 1 - Events and Their Probabilities

Expectation of Random Variables

Probability. Paul Schrimpf. January 23, UBC Economics 326. Probability. Paul Schrimpf. Definitions. Properties. Random variables.

SDS 321: Introduction to Probability and Statistics

Chapter 2. Continuous random variables

IAM 530 ELEMENTS OF PROBABILITY AND STATISTICS LECTURE 3-RANDOM VARIABLES

1 Probability theory. 2 Random variables and probability theory.

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.

Continuous Random Variables

ACM 116: Lectures 3 4

Statistics and Econometrics I

Lecture 2: Review of Probability

Random Variables and Their Distributions

conditional cdf, conditional pdf, total probability theorem?

2 Functions of random variables

Formulas for probability theory and linear models SF2941

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

1 Random variables and distributions

6.041/6.431 Fall 2010 Quiz 2 Solutions

Brief Review of Probability

Discrete Probability Refresher

Random Variables. Statistics 110. Summer Copyright c 2006 by Mark E. Irwin

SUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416)

ECE353: Probability and Random Processes. Lecture 7 -Continuous Random Variable

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

MAS113 Introduction to Probability and Statistics. Proofs of theorems

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 8 10/1/2008 CONTINUOUS RANDOM VARIABLES

Exam P Review Sheet. for a > 0. ln(a) i=0 ari = a. (1 r) 2. (Note that the A i s form a partition)

Review of Probability. CS1538: Introduction to Simulations

Probability. Paul Schrimpf. January 23, Definitions 2. 2 Properties 3

CME 106: Review Probability theory

MULTIVARIATE PROBABILITY DISTRIBUTIONS

Bivariate distributions

More than one variable

Fundamental Tools - Probability Theory II

ENGG2430A-Homework 2

Appendix A : Introduction to Probability and stochastic processes

The Multivariate Normal Distribution. In this case according to our theorem

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Continuous distributions

1 Review of Probability

LIST OF FORMULAS FOR STK1100 AND STK1110

Chapter 2: Random Variables

Lecture 1: August 28

STAT/MATH 395 A - PROBABILITY II UW Winter Quarter Moment functions. x r p X (x) (1) E[X r ] = x r f X (x) dx (2) (x E[X]) r p X (x) (3)

Probability Theory and Statistics. Peter Jochumzen

1: PROBABILITY REVIEW

Probability Review. Yutian Li. January 18, Stanford University. Yutian Li (Stanford University) Probability Review January 18, / 27

Jointly Distributed Random Variables

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Elements of Probability Theory

Continuous Random Variables and Continuous Distributions

Notes for Math 324, Part 19

Review: mostly probability and some statistics

Lecture 3: Random variables, distributions, and transformations

MATH4210 Financial Mathematics ( ) Tutorial 7

Bivariate Distributions

REVIEW OF MAIN CONCEPTS AND FORMULAS A B = Ā B. Pr(A B C) = Pr(A) Pr(A B C) =Pr(A) Pr(B A) Pr(C A B)

STAT/MATH 395 PROBABILITY II

Expectation, Variance and Standard Deviation for Continuous Random Variables Class 6, Jeremy Orloff and Jonathan Bloom

Preliminary Statistics. Lecture 3: Probability Models and Distributions

3 Multiple Discrete Random Variables

01 Probability Theory and Statistics Review

EXAM. Exam #1. Math 3342 Summer II, July 21, 2000 ANSWERS

Outline PMF, CDF and PDF Mean, Variance and Percentiles Some Common Distributions. Week 5 Random Variables and Their Distributions

Week 12-13: Discrete Probability

Preliminary Statistics Lecture 3: Probability Models and Distributions (Outline) prelimsoas.webs.com

18.175: Lecture 8 Weak laws and moment-generating/characteristic functions

If g is also continuous and strictly increasing on J, we may apply the strictly increasing inverse function g 1 to this inequality to get

Lecture 4 : Random variable and expectation

Transcription:

In probabilistic models, a random variable is a variable whose possible values are numerical outcomes of a random phenomenon. As a function or a map, it maps from an element (or an outcome) of a sample space to real values. The underlying probability over sample space makes it possible to define its cumulative distribution, which allows us to calculate quantities such as its expected value and variance, and the moments of its distribution. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R X : Ω R : ω X(ω) is called a random variable (r.v.), and usually determined by an experiment. We conventionally denote random variables by uppercase letters X, Y, Z, U, V, etc., from the end-of-alphabet letters. In particular, a discrete random variable is a random variable that can take values on a finite set {a 1, a,..., a n } of real numbers (usually integers), or on a countably infinite set {a 1, a, a 3,...}. The statement such as X = a i is an event since is a subset of a sample space Ω. {ω : X(ω) = a i } Frequency function. We can consider the probability of the event {X = a i }, denoted by P (X = a i ). The function p(a i ) := P (X = a i ) over the possible values of X, say a 1, a,..., is called a frequency function, or a probability mass function. The frequency function p must satisfy p(a i ) = 1, i where the sum is over the possible values of X. The frequency function will completely describe the probabilistic nature of random variable. Joint distributions of discrete random variables. When two discrete random variables X and Y are obtained in the same experiment, we can define their joint frequency function by p(a i, b j ) = P (X = a i, Y = b j ) where a 1, a,... and b 1, b,... are the possible values of X and Y, respectively. The marginal frequency function of X, denoted by p X, can be calculated by p X (a i ) = P (X = a i ) = j p(a i, b j ), where the sum is over the possible values b 1, b,... of Y. Similarly, the marginal frequency function p Y (b j ) = i p(a i, b j ) of Y is given by summing over the possible values a 1, a,... of X. Example 1. An experiment consists of throwing a fair coin three times. Let X be the number of heads, and let Y be the number of heads before the first tail. Page 1 Probability and Statistics I/September, 17

(a) List the sample space Ω. (b) Describe the events {X = }, {Y = }, and {X =, Y = }. (c) Find the frequency function p for X and Y. And compute the joint frequency p(, ). Solution. (a) Ω = {HHH, HHT, HT H, T HH, HT T, T HT, T T H, T T T } (b) {X = } = {T T T }, {Y = } = {T T T, T HH, T HT, T T H}, and {X =, Y = } = {T T T }, (c) p X () = 1 8 ; p X(1) = 3 8 ; p X() = 3 8 ; p X(3) = 1 8. p y () = 1 ; p y(1) = 1 4 ; p y() = 1 8 ; p y(3) = 1 8. p(, ) = P (X =, Y = ) = 1 8. Cumulative distribution function. Another useful function is the cumulative distribution function (cdf), and it is defined by F (x) = P (X x), < x <. The cdf of a discrete r.v. is a nondecreasing step function. It jumps wherever p(x) >, and the jump at a i is p(a i ). cdf s are usually denoted by uppercase letters, while frequency functions are usually denoted by lowercase letters. Independent random variables. Let X and Y be discrete random variables with joint frequency function p(x, y). Then X and Y are said to be independent, if they satisfy for all possible values of (x, y). p(x, y) = p X (x)p Y (y) Example. Continue the same experiment of throwing a fair coin three times. Let X be the number of heads, and let Y be the number of heads before the first tail. (a) Find the cdf F (x) for X at x = 1,, 1,,.5, 3, 4.5. (b) Are X and Y independent? Solution. (a) F ( 1) =, F () = 1, F (1) = 1, F () = F (.5) = 7, and F (3) = F (4.5) = 1. 8 8 (b) Since p X () = 1 8, p Y () = 1, and p(, ) = 1 8, we find that p(, ) p X()p Y (). Thus, X and Y are not independent. Page Probability and Statistics I/September, 17

Continuous random variables. A continuous random variable is a random variable whose possible values are real values such as 78.6, 5.7, 1.4, and so on. Examples of continuous random variables include temperature, height, diameter of metal cylinder, etc. In what follows, a random variable means a continuous random variable, unless it is specifically said to be discrete. The probability distribution of a random variable X specifies how its values are distributed over the real numbers. This is completely characterized by the cumulative distribution function (cdf). The cdf F (t) := P (X t). represents the probability that the random variable X is less than or equal to t. Then we say that the random variable X is distributed as F (t). Properties of cdf. The cdf F (t) must satisfy the following properties: (a) The cdf F (t) is a positive function. F (t) for all t. (b) The cdf F (t) increases monotonically. F (s) F (t) whenever s < t. (c) The cdf F (t) must tend to zero as t goes to negative infinity, and must tend to one as t goes to positive infinity +. lim F (t) = t lim F (t) = 1 t + Probability density function (pdf). It is often the case that the probability that the random variable X takes a value in a particular range is given by the area under a curve over that range of values. This curve is called the probability density function (pdf) of the random variable X, denoted by f(x). Thus, the probability that a X b can be expressed by P (a X b) = The pdf f(x) must satisfy the following properties: b a f(x)dx. (a) The pdf f(x) is a positive function [that is, f(x) ]. (b) The area under the curve of f(x) (and above the x-axis) is one [that is, f(x)dx = 1]. Relation between cdf and pdf. The pdf f(x) is related to the cdf F (t) via F (t) = P (X t) = t f(x)dx. Page 3 Probability and Statistics I/September, 17

This implies that such cdf F (t) is a continuous function, and that the pdf f(x) is the derivative of the cdf F (x) if f is continuous at x, that is, f(x) = df (x) dx. Example 3. A random variable X is called a uniform random variable on [a, b], when X takes any real number in the interval [a, b] equally likely. Then the pdf of X is given by { 1/(b a) if a x b; f(x) = otherwise [that is, if x < a or b < x]. Find the cdf F (t). Solution. F (t) = t if x < a; t a f(x)dx = if a x b; b a 1 if x > b. Quantile function. Given a cdf F (x), we can define the quantile function Q(u) = inf{x : F (x) u} for u (, 1). In particular, if F is continuous and strictly increasing, then we have Q(u) = F 1 (u). When F (x) is the cdf of a random variable X, and F (x) is continuous, we can find x p = Q(p) P (X x p ) = p for every p (, 1). The value x 1/ = Q( 1) is called the median of F, and the values x 1/4 = Q( 1) 4 and x 3/4 = Q( 3 ) are called the lower quartile and the upper quartile, respectively. 4 Example 4. Compute the quantile function Q(u) for a uniform random variable X on [a, b]. Find the median of distribution. Solution. Restricting the domain [a, b] for F, we find that F (x) = x a is strictly increasing b a function from (a, b) to (, 1), and therefore, that F (x) has the inverse function Q(u). By solving u = x a in terms of x, we obtain Q(u) = a + (b a)u for < u < 1. The median is given by b a Q ( ) 1 = a+b. Joint density function. Consider a pair (X, Y ) of random variables. A joint density function f(x, y) is a nonnegative function satisfying f(x, y) dy dx = 1, and is used to compute probabilities constructed from the random variables X and Y simultaneously by b [ d ] P (a X b, c Y d) = f(x, y) dy dx Page 4 Probability and Statistics I/September, 17 a c

In particular, F (u, v) = P (X u, Y v) = is called the joint distribution function. u v f(x, y) dy dx. Marginal densities and independence. Given the joint density function f(x, y), the distribution for each of X and Y is called the marginal distribution, and the marginal density functions of X and Y, denoted by f X (x) and f Y (y), are given respectively by f X (x) = f(x, y) dy and f Y (y) = f(x, y) dx. If the joint density function f(x, y) for continuous random variables X and Y is expressed in the form f(x, y) = f X (x)f Y (y) for all x, y, then X and Y are said to be independent. Example 5. Consider the following joint density function for random variables X and Y : { λ e λy if x y; f(x, y) = (.1) otherwise. (a) Find P (X 1, Y 1). (b) Find the marginal density functions f X (x) and f Y (y). (c) Are X and Y independent? Solution. For (a) we can calculate it as follows. P (X 1, Y 1) = = 1 1 y λ e λy dxdy = 1 λ e λy ydy λe λy dy [ λye λy] 1 = [ e λy] 1 λe λ = 1 e λ λe λ Continue the solution. (b) We obtain f X (x) = f Y (y) = for x and y. x y λ e λy dy = [ λe λy] y= y=x = λe λx λ e λy dx = [ λ e λy x ] x=y x= = λ ye λy (c) Since f(x, y) f X (x)f Y (y), they are not independent. Conditional density functions. Suppose that two random variables X and Y has a joint density function f(x, y). If f X (x) >, then we can define the conditional density function f Y X (y x) given X = x by f(x, y) f Y X (y x) = f X (x). Page 5 Probability and Statistics I/September, 17

Similarly we can define the conditional density function f X Y (x y) given Y = y by f X Y (x y) = f(x, y) f Y (y) if f Y (y) >. Then, clearly we have the following relation f(x, y) = f Y X (y x)f X (x) = f X Y (x y)f Y (y). Example 6. Find the conditional density functions f X Y (x y) and f Y X (y x) for the joint density function (.1) in the previous example. Solution. f X Y (x y) = f Y X (y x) = { f(x, y) λ e λy f Y (y) = = 1 if x y; λ ye λy y if x > y or x <. f(x, y) f X (x) = { if y < x; λ e λy λe λx = λe λ(y x) if y x. Independence of random variables. In theory we can consider a joint frequency function p(x, y, z) for three discrete random variables X, Y, and Z. If it satisfies p(x, y, z) = p X (x)p Y (y)p Z (z) then they are said to be independent. In the same manner, if a joint density function f(x, y, z) for continuous random variablesx, Y and Z is expressed in the form f(x, y, z) = f X (x)f Y (y)f Z (z) then they are independent. In general if X 1, X,..., X n are independent random variables then we can construct the joint density function by f(x 1, x,..., x n ) = f X1 (x 1 ) f X (x ) f Xn (x n ) Expectation. Let X be a discrete random variable whose possible values are a 1, a,..., and let p(x) is the frequency function of X. Then the expectation (expected value or mean) of the random variable X is given by E[X] = i a i p(a i ). We often denote the expected value E(X) of X by µ or µ X. For a function g, we can define the expectation of function of random variable by E[g(X)] = i g(a i )p(a i ). Page 6 Probability and Statistics I/September, 17

Variance. The variance of a random variable X, denoted by Var(X) or σ, is the expected value of the squared difference between the random variable and its expected value E(X), and can be defined as Var(X) := E[(X E(X)) ] = E[X ] (E[X]). The square-root Var(X) of the variance Var(X) is called the standard error (SE) (or standard deviation (SD)) of the random variable X. Example 7. A random variable X takes on values, 1, with the respective probabilities P (X = ) =., P (X = 1) =.5, P (X = ) =.3. Compute (a) E[X] (b) E[X ] (c) Var(X) and SD of X Solution. (a) E[X] = ()(.) + (1)(.5) + ()(.3) = 1.1 (b) E[X ] = () (.) + (1) (.5) + () (.3) = 1.7 (c) Var(X) = E[(X 1.1) ] = ( 1.1) (.) + (.1) (.5) + (.9) (.3) =.49. Also, using Var(X) = E[X ] (E[X]) we can calculate Var(X) = (1.7) (1.1) =.49. Then we obtain the SD of.49 =.7. Expectation for two variables. Suppose that we have two random variables X and Y, and that p(x, y) is their joint frequency function. Then the expectation of function g(x, Y ) of the two random variables X and Y is defined by E[g(X, Y )] = i,j g(a i, b j )p(a i, b j ), where the sum is over all the possible values of (X, Y ). Properties of expectation. One can think of the expectation E(X) as an operation on a random variable X which returns the average value for X. (a) Let a be a constant, and let X be a random variable having the frequency function p(x). Then we can show that E[a + X] = x (a + x)p(x) = a + x x p(x) = a + E[X]. Page 7 Probability and Statistics I/September, 17

(b) Let a and b be scalars, and let X and Y be random variables having the joint frequency function p(x, y) and the respective marginal density functions p X (x) and p Y (y). E[aX + by ] = x,y (ax + by)p(x, y) = a x x p X (x) + b y y p Y (y) = ae[x] + be[y ]. Linearity property of expectation. Let a and b 1,..., b n be scalars, and let X 1,..., X n be random variables. By applying the properties of expectation repeatedly, we can obtain [ ] [ n n ] [ n ] E a + b i X i = a + E b i X i = a + b 1 E[X 1 ] + E b i X i i=1 = = a + n b i E[X i ]. i=1 It is also useful to observe the above property as that of linear operator. i=1 Expectation for independent random variables. Suppose that X and Y are independent random variables. Then the joint frequency function p(x, y) of X and Y can be expressed as p(x, y) = p X (x)p Y (y). And the expectation of the function of the form g 1 (X) g (Y ) is given by E[g 1 (X)g (Y )] = g 1 (x)g (y) p(x, y) x,y [ ] [ ] = g 1 (x)p X (x) g (y)p Y (y) = E[g 1 (X)] E[g (Y )]. x y i= Covariance. Suppose that we have two random variables X and Y. Then the covariance of two random variables X and Y can be defined as Cov(X, Y ) := E((X µ x )(Y µ y )) = E(XY ) E(X) E(Y ), where µ x = E(X) and µ y = E(Y ). Then the correlation coefficient ρ = Cov(X, Y ) Var(X)Var(Y ) measures the strength of the dependence of the two random variables. Properties of variance and covariance. (a) If X and Y are independent, then Cov(X, Y ) = by observing that E[XY ] = E[X] E[Y ]. Page 8 Probability and Statistics I/September, 17

(b) In contrast to the expectation, the variance is not a linear operator. variables X and Y, we have For two random Var(X + Y ) = Var(X) + Var(Y ) + Cov(X, Y ). (.) Moreover, if X and Y are independent, by observing that Cov(X, Y ) = in (.), we obtain Var(X + Y ) = Var(X) + Var(Y ). In general, we have Var(X 1 + + X n ) = Var(X 1 ) + + Var(X n ). if X 1,..., X n are independent random variables. Variance and covariance under linear transformation. Let a and b be scalars (that is, real-valued constants), and let X be a random variable. Then the variance of ax + b is given by Var(aX + b) = a Var(X). Now let a 1, a, b 1 and b be scalars, and let X and Y be random variables. Then similarly the covariance of a 1 X + b 1 and a Y + b can be given by Cov(a 1 X + b 1, a Y + b ) = a 1 a Cov(X, Y ) Example 8. The joint frequency function p(x, y) of two discrete random variables, X and Y, is given by X 1 1 Y 1 1/ 1 1/4 1/4 (a) Find the marginal frequency function for X and Y. (b) Find E[X] and E[Y ]. (c) Find Cov(X, Y ). (d) Are X and Y independent? Solution. (a) p X ( 1) = 1 4 ; p X() = 1 ; p X(1) = 1 4. p Y ( 1) = 1 ; p X(1) = 1. (b) E[X] = ( 1) ( 1 4) + () ( 1 ) + (1) ( 1 4) = E[Y ] = ( 1) ( 1 ) + (1) ( 1 ) = (c) E[XY ] = ( 1)(1) ( 1 4) + ()( 1) ( 1 ) + (1)(1) ( 1 4) = Thus, we obtain Cov(X, Y ) = E[XY ] E[X]Y [Y ] =. Page 9 Probability and Statistics I/September, 17

(d) No, X and Y are not independent, because p( 1, 1) = p X ( 1)p Y ( 1) = 1 8. Expectation of continuous random variables. Let f(x) be the pdf of a continuous random variable X. Then we define the expectation of the random variable X by E[X] = x f(x) dx. Furthermore, we can define the expectation E[g(X)] of the function g(x) of random variable by E[g(X)] = g(x) f(x) dx, if g(x) is integrable with respect to f(x) dx, that is, g(x) f(x) dx <. Properties of expectation. When X is a discrete random variable, E[X] is considered as a linear operator. This remains true even if X is a continuous random variable. Thus, we have the following properties (without proof): (a) E[a + X] = a + E[X]. (b) E[aX + by ] = ae[x] + be[y ]. [ ] n n (c) E a + b i X i = a + b i E[X i ]. i=1 i=1 (d) If X and Y are independent then E[g 1 (X)g (Y )] = E[g 1 (X)] E[g (Y )]. Example 9. Let X be a uniform random variable on [a, b]. Compute E[X] and E[X ], and find Var(X). Solution. b ( ) ( ) [ 1 1 x E[X] = x dx = a b a b a = a + b E[X ] = b = a + ab + b 3 Var(X) = a + ab + b a ( ) ( ) [ 1 1 x x 3 dx = b a b a 3 3 ( ) a + b = ] b a ] b (b a) 1 a Page 1 Probability and Statistics I/September, 17

Moment generating function. Let X be a random variable. Then the moment generating function (mgf ) of X is given by M(t) = E [ e tx]. The mgf M(t) is a function of t defined on some open interval (c, c 1 ) with c < < c 1. In particular when X is a continuous random variable having the pdf f(x), the mgf M(t) can be expressed as M(t) = e tx f(x) dx. Properties of moment generating function. (a) The most significant property of moment generating function is that the moment generating function uniquely determines the distribution. (b) Let a and b be constants, and let M X (t) be the mgf of a random variable X. Then the mgf of the random variable Y = a + bx can be given as follows. M Y (t) = E [ e ty ] = E [ e t(a+bx)] = e at E [ e (bt)x] = e at M X (bt). (c) Let X and Y be independent random variables having the respective mgf s M X (t) and M Y (t). We can obtain the mgf M Z (t) of the sum Z = X + Y of random variables as follows. M Z (t) = E [ e tz] = E [ e t(x+y )] = E [ e tx e ty ] = E [ e tx] E [ e ty ] = M X (t) M Y (t). Properties of moment generating function, continued. (d) When t =, it clearly follows that M() = 1. Now by differentiating M(t) r times, we obtain M (r) (t) = dr dt E [ e tx] [ ] d r = E r dt r etx = E [ X r e tx]. In particular when t =, M (r) () generates the r-th moment of X as follows. M (r) () = E[X r ], r = 1,, 3,... Example 1. Find the mgf M(t) for a uniform random variable X on [a, b]. Then show that (a) lim t M(t) = 1 and (b) lim t M (t) = a+b. Solution. M(t) = b a ( ) 1 e tx dx = b a b ( 1 a ( 1 b a b a) dx = 1 if t = ; ) [ ] x=b e tx = etb e ta if t. t t(b a) e tb e ta lim M(t) = lim t t t(b a) = lim be tb ae ta = 1 by l Hospital s rule. t b a lim M (be tb ae ta )t (e tb e ta ) (t) = lim t t t (b a) = lim t b e tb a e ta (b a) = a + b x=a again, by l Hospital s rule. Page 11 Probability and Statistics I/September, 17

Page 1 Probability and Statistics I/September, 17

Exercises Exercises Problem 1. Let p(k), k = 1,, 1, be the frequency function for random variable X. Suppose that p() = 1, and that p( 1) and p(1) are unknown. 4 (a) Show that E[X ] does not depend on the unknown values p( 1) and p(1). (b) If E[X] = 1, then find the values p( 1) and p(1). 4 Problem. The joint frequency function p(x, y) of two discrete random variables, X and Y, is given by X 1 3 Y 1 c 3c c (a) Find the constant c. (b) Find E[X] and E[XY ]. (c) Are X and Y independent? c c c Problem 3. A pair (X, Y ) of discrete random variables has the joint frequency function (a) Find P (X + Y = 3). p(x, y) = xy, x = 1,, 3 and y = 1,. 18 (b) Find the marginal frequency function for X and Y. (c) Find E[Y ] and Var(Y ). (d) Are X and Y independent? Justify your answer. Problem 4. (a) Determine the constant c so that p(x) is a frequency function if p(x) = cx, x = 1,, 3, 4, 5, 6. (b) Similarly find c if p(x) = c ( 3) x, x = 1,, 3, 4,.... Problem 5. Let X and Y be independent random variables. (a) Show that Var(aX) = a Var(X). (b) If E[X] = 1, E[Y ] = and Var(X) = 4, Var(Y ) = 9 then find the mean and the variance of Z = 3X Y. Page 13 Probability and Statistics I/September, 17

Exercises Problem 6. Suppose that X has the density function f(x) = cx for x 1 and f(x) = otherwise. (a) Find c. (b) Find the cdf. (c) What is P (.1 X <.5)? Problem 7. Let X be a random variable with the pdf { x if x 1; f(x) = otherwise. (a) Find E[X]. (b) Find E[X ] and Var(X). Problem 8. Let F (x) = 1 exp( αx β ) for x, α > and β >, and F (x) = for x <. Show that F is a cdf, and find the corresponding density. Problem 9. Suppose that random variables X and Y have the joint density { 6 7 f(x, y) = (x + y) if x 1 and y 1; otherwise. (a) Find the marginal density functions of X and Y. (b) Find the conditional density functions of X given Y = y and of Y given X = x. (c) Find E[X] and E[Y ]. Problem 1. Discuss the following properties. (a) Prove the property of variance: Var(aX + b) = a Var(X). (b) Suppose that E[X] = µ and Var(X) = σ. Then Z = (X µ)/σ is said to be standardized, in which σ is often referred as standard error. Show that E[Z] = and Var(Z) = 1. Problem 11. Recall that the expectation is an linear operator. By using the definitions of variance and covariance in terms of the expectation, show that (a) Var(X + Y ) = Var(X) + Var(Y ) + Cov(X, Y ), and (b) Var(X Y ) = Var(X) + Var(Y ) Cov(X, Y ). Page 14 Probability and Statistics I/September, 17

Exercises Problem 1. If X and Y are independent random variables with equal variances, find Cov(X + Y, X Y ). Problem 13. Suppose that X and Y are independent random variables with their variances σ x = Var(X) and σ y = Var(Y ). Let Z = Y X. (a) Find expressions for the covariance of X and Z in terms of σ x and σ y. (b) Find the correlation coefficient ρ of X and Z in terms of σ x and σ y. Problem 14. Let X and Y be independent random variables, and let α and β be scalars. Find an expression for the mgf M Z (t) of Z = αx + βy in terms of the mgf s M X (t) and M Y (t) of X and Y. Page 15 Probability and Statistics I/September, 17

Optional problems Optional problems Problem 15. The Cauchy distribution function is defined by F (x) = 1 + 1 π tan 1 (x), < x <. (a) Show that this is a cdf. (b) Find the density function. (c) Suppose that a random variable X has the Cauchy distribution. Can you find E[X]? Why or why not? Problem 16. Let f(x) = (1+αx)/ for 1 x 1 and f(x) = otherwise, where 1 α 1. (a) Show that f is a density. (b) Find the cdf F (x), and sketch the graph of F (x) for 1 α <, α =, and < α 1. (c) Find the formulas for the quartile function Q(u). Hint: Consider the following three cases separately: 1 α <, α =, and < α 1. Page 16 Probability and Statistics I/September, 17

Answers to exercises Answers to exercises Problem 1. (a) E[X ] = p( 1) + p(1) = 1 p() = 3 4 (b) E[X] = p( 1) + p(1) = 1 4. Together with p( 1) + p(1) = 3 4, we obtain p( 1) = 1 4 and p(1) = 1. Problem. (a) 3 x=1 y=1 p(x, y) = 1c = 1 implies that c = 1 (b) E[X] = (1) ( ( 1 5) + () ( 5) + (3) 5 E[Y ] = (1) ( ( 3 5) + () ) 5 = 7 5 E[XY ] = (1)(1) ( ) ( 1 1 +()(1) 3 1 ) = 11 5 1. ) ( +(3)(1) ) ( 1 +(1)() 1 ) ( 1 +()() 1 ) ( 1 +(3)() ) 1 = 31 1 (c) X and Y are not independent because p(1, 1) = 1 p 1 X(1)p Y (1) = 3. Or, you can find 5 it by calculating Cov(X, Y ) = E[XY ] E[X]E[Y ] = 1 5 Problem 3. (a) P (X + Y = 3) = p(1, ) + p(, 1) = 9 (b) p X (y) = x 6 p Y (y) = y 3 for x = 1,, 3. for y = 1,. (c) E[Y ] = (1) ( ( 1 3) + () ) 3 = 5 ; E[Y ] = (1) ( ) ( 1 3 3 + () 3) = 3 Var(Y ) = E[Y ] (E[Y ]) = 9 (d) Yes, since the joint frequency function satisfies p(x, y) = p X (x)p Y (y) for all x = 1,, 3 and y = 1,. Problem 4. (a) 6 6(6+1) x=1 p(x) = c = 1c = 1 Thus, we obtain c = 1 1 (b) (/3) x=1 p(x) = c = c = 1 Thus, we obtain c = 1 1 (/3) Problem 5. (a) Var(aX) = E[(aX E[aX]) ] = E[(aX ae[x]) ] = E[a (X E[X]) ] = a E[(X E[X]) ] = a Var(X). Page 17 Probability and Statistics I/September, 17

Answers to exercises (b) E[Z] = E[3X Y ] = (3)(1) ()() = 1 Var(Z) = Var(3X + ( )Y ) = Var(3X) + Var(( )Y ) = (3) (4) + ( ) (9) = 7 Problem 6. 1 (a) cx dx = c = 1. Thus, we have c = 3. 3 if x < ; (b) F (t) = t 3 if x 1; 1 if x > 1. (c) P (.1 X <.5) = Problem 7. (a) E[X] = 1 1.5.1 [ x 3 x(x) dx = 3 3x dx = [ x 3].5.1 =.14 ] 1 [ x (b) E[X ] = x 4 (x) dx = 1 ( ) 3 = 1. 18 = 3 ] 1 = 1 Then we obtain Var(X) = E[X ] (E[X]) = Problem 8. F (x) is nonnegative and increasing for x, and satisfies F () =, and lim x F (x) = 1. Thus, F (x) is cdf. The pdf is given by for x. f(x) = d dx F (x) = αβxβ 1 exp( αx β ) Problem 9. (a) f X (x) = 1 f Y (y) = 1 6 (x + 7 y) dy = 7 (3x + 3x + 1), x 1; 6 (x + 7 y) dx = 7 (3y + 3y + 1), y 1. (b) f X Y (x y) = 3(x+y) 3y +3y+1, x 1; f Y X (y x) = 3(x+y), y 1; 3x +3x+1 (c) E[X] = 1 (x) ( ) 7 (3x + 3x + 1)dx = 9 = E[Y ]. 14 Problem 1. Page 18 Probability and Statistics I/September, 17

Answers to exercises (a) Since E[aX + b] = ae[x] + b, by applying the definition of variance we obtain Var(aX + b) = E[(aX + b E[aX + b]) ] = E[a (X E[X]) ] = a E[(X E[X]) ] = a Var(X) (b) Since Z = ( ( 1 σ) X µ σ) is a linear transformation, we obtain ( ) 1 ( µ ) E[Z] = E[X] = σ σ ( ) 1 Var(Z) = Var(X) = 1 σ Problem 11. Let µ x = E[X] and µ y = E[Y ]. Then we show (a) as follows: Var(X + Y ) = E[(X + Y (µ x + µ y )) ] = E[((X µ x ) + (Y µ y )) ] We can similarly verify (b). = E[(X µ x ) + (X µ x )(Y µ y ) + (Y µ y ) ] = E[(X µ x ) ] + E[(X µ x )(Y µ y )] + E[(Y µ y ) ] = Var(X) + Cov(X, Y ) + Var(Y ) Problem 1. Let µ x = E[X] and µ y = E[Y ]. By applying the definition of variance and covariance, we obtain Cov(X + Y, X Y ) = E[(X + Y (µ X + µ y ))(X Y (µ x µ y ))] = E[(X µ X ) (Y µ y ) ] = E[(X µ X ) ] E[(Y µ y ) ] = Var(X) Var(Y ) = where the last equality comes from the assumption of equal variances. Problem 13. (a) Let µ x = E[X] and µ y = E[Y ]. Since Cov(X, Y ) =, we obtain Cov(X, Z) = E[(X µ x )((Y X) (µ y µ x ))] = E[(X µ x )(Y µ y ) (X µ x ) ] (b) We can calculate = Cov(X, Y ) Var(X) = σ x Then we obtain Var(Z) = Var(Y X) = Var(Y ) + Var( X) = σy + σx Cov(X, Z) σ ρ = = x Var(X)Var(Z) σ x + σy Problem 14. By the property of expectation for two independent random variables, we obtain M Z (t) = E[e t(αx+βy ) ] = E[e (tα)x ] E[e (tβ)y ] = M X (αt) M Y (βt) Page 19 Probability and Statistics I/September, 17