Module 3. Function of a Random Variable and its distribution

Size: px
Start display at page:

Download "Module 3. Function of a Random Variable and its distribution"

Transcription

1 Module 3 Function of a Random Variable and its distribution 1. Function of a Random Variable Let Ω, F, be a probability space and let be random variable defined on Ω, F,. Further let h: R R be a given function and let : Ω R be a function of random variable, defined by = h, Ω. In many situations it may be of interest to study the probabilistic properties of, which is a function of random variable. Since the variable takes values in R, to study the probabilistic properties of, it is necessary that F, B, i.e., is a random variable. Throughout, for a positive integer, R will denote the -dimensional Euclidean space and B will denote the Borel sigma-field in R. Definition 1.1 Let and be positive integers. A function h: R R is said to be a Borel function if h B, B. The following lemma will be useful in deriving conditions on the function h: R R so that : Ω R, defined by = h, Ω, is a random variable. Recall that, for a function Ψ: and, Ψ = : Ψ. Lemma 1.1 Let : Ω R and h: R R be given functions. Define : Ω R by = h, Ω. Then, for any R, = h. Proof. Fix R. Note that h = R: h. Clearly Therefore, h h. = : = : h = : h = h 1

2 Theorem 1.1 Let be a random variable defined on a probability space Ω, F, and let h: R R be a Borel function. Then the function : R, defined by = h,, is a random variable. Proof. Fix B. Since h is a Borel function, we have h B. Now using the fact that is a random variable it follows that This proves the result. Remark 1.1 = h F. (i) Let h: R R be a continuous function. According to a standard result in calculus inverse image of any open interval,, <, under continuous function h is a countable union of disjoint open intervals. Since B contains all open intervals and is closed under countable unions it follows that h, B, whenever <. Now on employing the arguments similar to the one used in proving Theorem 1.1, Module 2 (also see Theorem 1.2, Module 2) we conclude that h B, B. It follows that any continuous function h: R R is a Borel function and thus, in view of Theorem 1.1, any continuous function of a random variable is a random variable. In particular if is a random variable then,, max, 0, sin and cos are random variables. (ii) Let h: R R be a strictly monotone function. Then, for <, h, is a countable union of intervals and therefore h, B, i.e., h is a Borel function. It follows that if is a random variable and if h: R R is strictly monotone then h is a random variable. A random variable takes values in various Borel sets according to some probability law called the probability distribution of random variable. Clearly the probability distribution of a random variable of absolutely continuous/discrete type is described by its distribution function (d.f.) and/or by its probability density function/probability mass function (p.d.f/p.m.f.). For a given Borel function h: R R, in the following section, we will derive probability distribution of h using the probability distribution of random variable. 2

3 2. Probability Distribution of a Function of a Random Variable In our future discussions when we refer to a random variable, unless otherwise stated, it will be either of discrete type or of absolutely continuous type. The probability distribution of a discrete type random variable will be referred to as a discrete (probability) distribution and the probability distribution of a random variable of absolutely continuous type will be referred to as an absolutely continuous (probability) distribution. The following theorem deals with discrete probability distributions. Theorem 2.1 Let be a random variable of discrete type with support and p.m.f.. Let h: R R be a Borel function and let : Ω R be defined by = h, Ω.Then is a random variable of discrete type with support = h: and p.m.f. =, if 0, otherwise where = : h =. =, if 0, otherwise, Proof. Since h is a Borel function, using Theorem 1.1, it follows that is a random variable. Also is of discrete implies that is countable which further implies that is countable. Fix, so that = h for some. Then = = Ω: = : h = h = h = h = =, and = Ω: : h = h =. Therefore, = > 0, (since ), and = 1. 3

4 It follows that is countable, = > 0, and = 1, i. e., is a discrete type random variable with support. Moreover, for, Hence the result follows. = = Ω: h = = = = =. The following corollary is an immediate consequence of the above theorem. Corollary 2.1 Under the notation and assumptions of Theorem 2.1, suppose that h: R R is one-one with inverse function h : R, where = h: R. Then is a discrete type random variable with support = h: and p.m.f. = h, if 0, otherwise = = h, if 0, otherwise. Example 2.1 Let be a random variable with p.m.f. 1 7, if 2, 1, 0, 1 = 3., if 2, , otherwise Show that = is a random variable. Find its p.m.f. and distribution function. Solution. Since h =, R, is a continuous function and is a random variable, using Remark 1.1 (i) it follows that = h = is a random variable. Clearly = 2, 1, 0, 1, 2, 3 and = 0, 1, 4, 9. Moreover, 4

5 = 0 = = 0 = = 0 = 1 7, = 1 = = 1 = 1, 1 = = 2 7, = 4 = = 4 = 2, 2 = = 5 14, and = 9 = = 9 = 3, 3 = = 3 14 Therefore the p.m.f. of is and the distribution function of is 1, if = 0 7 2, if = 1 7 = 5,, if = , if = , otherwise 0, if < 0 1, if 0 < =, if 1 < , if 4 < , if z 9 Example 2.2 Let be a random variable with p.m.f., if ± 1, ±2,, ±50 =. 0, otherwise Show that = is a random variable. Find its p.m.f., and distribution function. 5

6 Solution. As h =, R, is a continuous function and is a random variable, using Remark 1.1 (i), = is a random variable. We have = ± 1, ±2,, ±50 and = 1, 2,, 50. Moreover, for, Therefore the p.m.f. of is = = = =, = = 1275 and the distribution function of is Example 2.3 Let be a random variable with p.m.f., if 1, 2,, 50 = 1275, 0, otherwise 0, if < 1, if 1 < 2 =., if < + 1, = 2, 3,,49 1, if 50 = 1, if 0, 1,,, 0, otherwise where is a positive integer and 0,1. Show that = is a random variable. Find its p.m.f. and distribution function. Solution. Note that = = 0, 1,, and h =, R, is a continuous function. Therefore = is a random variable. For = = = = 1 = 1. Thus the p.m.f. of is = 1, if 0, 1,,, 0, otherwise 6

7 and the distribution function of is 0, if < 0, if 0 < 1 = 1., if < + 1, = 1,2,, 1 1, if The following theorem deals with probability distribution of absolutely continuous type random variables. Theorem 2.2 Let be a random variable of absolutely continuous type with p.d.f. and support. Let,,,, be open intervals inr such that =, if and =. Let h: R R be a Borel function such that, on each = 1,,, h: R is strictly monotone and continuously differentiable with inverse function h. Let h = h: so that h = 1,, is an open interval in R. Then the random variable = h is of absolutely continuous type with p.d.f. = h h. Proof. We will provide an outline of the proof which may not be rigorous. Let be the distribution function of. For R and Δ > 0, + Δ = Δ < h + Δ Δ = < h + Δ, Δ Fix 1,,. First suppose that h is strictly decreasing on. Note that = h h and h is an open interval. Thus, for belongings to the exterior of h and sufficiently small Δ > 0, we have < h + Δ, = 0. Also, for h and sufficiently small Δ > 0, < h + Δ, = h + Δ < h. 7

8 Thus, for all R, we have < h + Δ, Δ = h + Δ < h Δ = 1 Δ Similarly if h is strictly increasing on then, for all R, we have < h + Δ, Δ h h. (2.1) = h < h + Δ Δ = 1 Δ h h. (2.2) Note that if h is strictly decreasing (increasing) on then h < >0 on. Now on combining (2.1) and (2.2) we get, for all R, < h + Δ, Δ + Δ Δ Similarly one can show that, for all R, h h, h h. +Δ lim Δ Δ = h h. 2.3 It follows that the distribution function of is differentiable everywhere on R except possibly at a finite number of points (on boundaries of intervals h,, h of ). Now the result follows from Remark 4.2 (vii) of Module 2 and using (2.3). The following corollary to the above theorem is immediate. 8

9 Corollary 2.2 Let be a random variable of absolutely continuous type with p.d.f. and support. Suppose that is a finite union of disjoint open intervals in R and let h: R R be a Borel function such that h is differentiable and strictly monotone on (i.e., either h < 0, or h > 0, ). Let = h:. Then = h is a random variable of absolutely continuous type with p.d.f. = h h, if 0, otherwise It may be worth mentioning here that, in view of Remark 4.2 (vii) of Module 2, Theorem 2.2 and Corollary 2.2 can be applied even in situations where the function h is differentiable everywhere on except possibly at a finite number of points. Example 2.4 Let be random variable with p.d.f. and let = =, if > 0 0, otherwise, (i) (ii) (iii) Show that is a random variable of absolutely continuous type; Find the distribution function of and hence find its p.d.f.; Find the p.d.f. of directly (i.e., without finding the distribution function of ). Solution. (i) and (iii). Clearly = is a random variable (being a continuous function of random variable ). We have = = 0,. Also h =,, is strictly increasing on with inverse function h =,. Using Corollary 2.1 it follows that = is a random variable of absolutely continuous type with p.d.f. =, if > 0 0, otherwise =, if > 0. 0, otherwise (ii) We have =, R. Clearly, for < 0, = 0. For 0, 9

10 = = Therefore the distribution function of is = = 1. 0, if < 0 = 1, if 0. Clearly is differentiable everywhere except at = 0. Therefore, using Remark 4.2 (vii) of Module 2, we conclude that the random variable is of absolutely continuous type with p.d.f. =, if 0. At = 0 we may assign any arbitrary non- negative value to 0. Thus a p.d.f. of is Example 2.5 Let be a random variable with p.d.f. and let = =, if > , otherwise =, if 1 < < 1, if 1 < 2 0, otherwise, (i) (ii) (iii) Show that is a random variable of absolutely continuous type; Find the distribution function of and hence find its p.d.f; Find the p.d.f. of directly (i.e., without finding the distribution function of ). 10

11 Solution. (i) and (iii). Clearly = is a random variable (being a continuous function of random variable ). We have = 1, 0 0, 2 =, say. Also h =,, is strictly decreasing in = 1, 0 with inverse function h = ; h =,, is strictly increasing in = 0, 2, with inverse function h = ; h = 0, 1 and h = 0, 4. Using Theorem 2.2 it follows that = is a random variable of absolutely continuous type with p.d.f. =, +, 1, if 0 < < 1 2 = 1, if 1 < < 4 6 0, otherwise (ii) We have =, R. Since 1, 2 = 1, we have 0, 4 = 1. Therefore, for < 0, = = 0 and, for 4, = = 1. For [0,4, we have = =, if 0 < 1 2 = , if 1 < 4 Therefore, the distribution function of is 0, if < 0, if 0 < 1 ) = 2 + 2, if 1 < 4 6 1, if 4 Clearly is differentiable everywhere except on set = 0, 1, 2. Also ) ) = =

12 Using Remark 4.2 (vii) of Module 2 it follows that the random variable is of absolutely continuous type with a p.d.f. 1, if 0 < < 1 2 = 1, if 1 < < 4 6 0, otherwise Note that a Borel function of a discrete type random variable is a random variable of discrete type (see Theorem 1.1). Theorem 2.2 provides sufficient conditions under which a Borel function of an absolutely continuous type random variable is of absolutely continuous type. The following example illustrates that, in general, a Borel function of an absolutely continuous type random variable may not be of absolutely continuous type. Example 2.6 Let be a random variable of absolutely continuous type with p.d.f. =, if > 0 0, otherwise, and let = [, where, for R, [ denotes the largest integer not exceeding. Show that is a random variable of discrete type and find its p.m.f.. Solution. For R, we have, =, [ + 1 B. It follows that is a random variable. Also = 0,. Since = 1, we have 0, 1, 2, = 1. Also, for 0, 1, 2,. = = < + 1 = = = 1 12

13 > 0. Consequently the random variable is of discrete type with support = 0, 1, 2, and p.m.f. = = = 1, if 0,1, 2,. 0, otherwise 3. Expectation and Moments of a Random Variable Suppose that is a discrete type random variable defined on a probability space Ω, F, associated with a random experiment. Let and denote, respectively, the support and the p.m.f. of. Suppose that the random experiment is repeated a large number of times. Let,, denote the frequency of the event = in the first repetitions of the random experiment. Then, according to the relative frequency approach to the probability, = = lim,. Note that represents the mean observed value (or expected value) of random variable in the first repetitions of the random experiment. Therefore, in line with axiomatic approach to probability, one may define the mean value (or expected value) of random variable as = lim = lim = = =, 13

14 provided the involved limits exist and the interchange of signs of summation and limit is admissible. A similar discussion can be provided for defining the expected value of an absolutely continuous type random variable, having p.d.f., as provided the integral is defined. =, The above discussion leads to the following definitions. Definition 3.1 (i) Let be a discrete type random variable with p.m.f. and support. We say that the expected value of (denoted by ) is finite and equals =, provided ) <. (ii) Let be an absolutely continuous type random variable with p.d.f.. We say that the expected value of (denoted by ) ) is finite and equals provided ) < ) = ), The following observations to above definitions are immediate. Remark 3.1 (i) Since ) ) = ) and ) ) = ), it follows that if the expected value of a random variable is finite, then ) <. 14

15 (ii) (iii) If is a random variable of discrete type with finite support, then ) <. Consequently the expected value of is finite. Suppose that is a random variable of absolutely continuous type with p.d.f. and support,, for some 0, ). Then Example 3.1 ) = ) ) = <. Consequently the expected value of is finite. Let be a random variable with p.m.f. ) =, if 1, 2, 3,. 0, otherwise Show that the expected value of is finite and find its value. Solution. We have = 1, 2, 3, and ) = 2 =, say. Clearly = > 0, = 1, 2, and By the ratio test, = < 1, as. 2 ) = <, and therefore the expected value of is finite. Moreover, 15

16 = = 2 = lim, where = On subtracting (3.2) from (3.1), we get = 2 = = = = = lim = 2 Example 3.2 Let be a random variable with p.m.f. Show that the expected of is not finite. Solution. We have = ±1, ±2, ±3, and 3 =, if ±1, ±2, ±3,. 0, otherwise ) = 6 π 1 = 16

17 Thus the expected value of is not finite. Example 3.3 Let be random variable with p.d.f. =, < <. 2 Show that the expected value of is finite and find its value. Solution. We have = = = 1. Thus the expected value of is finite and 2 = = = 0. 2 Example 3.4 Let be random variable with p.d.f. Show that the expected value of is not finite. Solution. We have = 1 1, < <

18 = = =. Therefore the expected value of is not finite. Theorem 3.1 Let be a random variable of absolutely continuous or discrete type with finite expected value. Then (i) = > < ; (ii) = >, provided 0 = 1; (iii) =, provided 0, ±1, ±2, = 1; (iv) =, provided 0, 1, 2, = 1. Proof. (i) Case I. is of absolutely continuous type = = + = + On changing the order of integration in the two integrals above, we get Case II. is of discrete type = + = < + >. 18

19 We will illustrate the idea of the proof by considering a special case where =,,, < < < < < 0 < < < < and lim =. Under the above situation > = > + > Also, = + = + = + = + = [ + [ = = < = < + < + < =

20 = = = = = Therefore, > < = = =. (ii) Suppose that 0 = 1. Then < = 0, 0, and therefore = > < = >. (iii) Suppose that 0, ±1, ±2,. Then, for (the set of integers) and 1, we have Therefore, > = and < = 1 > = > = =, 20

21 < = < = =, and the assertion follows on using (i). (iv) Suppose that 0, 1, 2, = 1. Then = 0, 1, 2, and the result follows from (iii). Theorem 3.2 (i) (ii) Let be a random variable of discrete type with support and p.m.f.. Let h: R R be a Borel function and let = h. Then = h, provided it is finite. Let be a random variable of absolutely continuous type with p.d.f.. Let h: R R be a Borel function and let = h. Then provided it is finite. = h, Proof. (i) By Theorem 2.1, = h is a random variable of discrete type with support = h: and p.m.f. =, = = = if, 0, otherwise where = : h =,, so that : forms a partition of =, if, and =. Therefore, 21

22 = = = = = = = h = since for, = h = h = since =, if = h = since = = h. (ii) Define = : h >, 0, and = : h <, 0. For simplicity we assume that, for every 0 and 0, and are intervals. Then, using Theorem 3.1 (i), = > < = =, on interchanging the order of integration in the above two integrals and using the following two immediate observations: (a) 0,, and 0, h; (b), 0, and h, 0. Therefore, = h + h 22

23 Remark 3.2 = h = h, since = and = : h = 0. Recall that probability density function of absolutely continuous type random variable is not unique. However the distribution function of any random variable is unique. Theorem 3.1 (i) implies that the expected value = = 1, of an absolutely continuous type random variable is unique (i.e., it does not depend on the version of p.d.f. used) although the probability density function may not be unique. Some special kind of expectations which will be frequently used are defined below. Definition 3.2 Let be a random variable defined on some probability space. (i) (ii) (iii) (iv) (v) (vi) =, provided it is finite, is called the mean of the (distribution of) random variable ; For 1, 2, =, provided it is finite, is called the r-th moment of the (distribution of) random variable ; For 1, 2,, provided it is finite, is called the r-th absolute moment of the (distribution of) random variable ; For 1, 2, =, provided it is finite, is called the r-th central moment of the (distribution of) random variable ; =, provided it is finite, is called the variance of the (distribution of) random variable. The variance of a random variable is denoted by Var. The quantity = = is called the standard deviation of the (distribution of) random variable. Suppose that the distribution function of a random variable X can be decomposed as 23

24 Theorem 3.3 = + 1, R, [0,1, where is a distribution function of a discrete type random variable (say ) and is a distribution function of an absolutely continuous type random variable (say ). Then, for a Borel function h: R R, the expectation of h is defined by h = h + 1 h, provided h, h and h are finite. Let be a random variable. (i) If h and h are Borel functions such that h h = 1, then h h, provided the involved expectations are finite; (ii) If, for real constants and with, = 1, then ; (iii) If 0 = 1 and = 0, then = 0 = 1; (iv) If is finite, then ; (v) For real constants and, + = +, provided the involved expectations are finite; (vi) If h,, h are Borel function then h = h, provided the involved expectations are finite. Proof. We will provide the proof for the situation when is of absolutely continuous type. The proof for the discrete case is analogous and is left as an exercise. Also assertions (iv)-(vi) follow directly from the definition of the expectation of a random variable and using elementary properties of integrals (sums/series). Therefore we will provide the proofs of only first three assertions. (i) Define = R: h h, = and =, if 0, otherwise. Then 0, R, = 0, = 0, = = + = = 1, 24

25 = = = 1 and, for any B, = since = 1 = since = 0 = =. It follows that is also a p.d.f. of with support =. The above discussion suggest that, without loss of generality, we may take = R: h h (otherwise replace by and by ). Then h h, R h = h h = h. (ii) (iii) Since = 1, as in (i), without loss of generality we may assume that [,. Then, R = =, i.e.,. Since 0 = 1, without loss of generality we may take [0,. Then, 0 = R: = 0 and therefore, for 1, 2,, 0 = = + = 25

26 1 = = 0, 1, 2, lim 1 = 0 1 = 0 since 1 > 0 = 0 = 0 = 0 > 0 = 1. Corollary 3.1 Let be random variable with finite first two moments and let =. Then, (i) Var = ; (ii) Var 0. Moreover, Var = 0 if, and only if, = = 1; (iii) (Cauchy Schwarz inequality); (iv) For real constants and, Var + = Var. Proof. (i) Note that = is a fixed real number. Therefore, using Theorem 3.3 (v)-(vi), we have Var = = 2 + = =. 26

27 (ii) Since 0 = Ω = 1, using Theorem 3.3 (i), we have Var = 0. Also, using Theorem 3.3 (iii), if Var = = 0 then = 0 = 1, i. e., = = 1. Conversely if = = 1, then = and =. Now using (i), we get Var = = 0. (iii) (iv) Follows from (i) and (ii). Let = +. Then = + (using Theorem 3.3 (v)) = and Var = = = = Var Example 3.5 Let be a random variable with p.d.f. 1 2, if 2 < < 1 =, if 0 < < 3 9 0, otherwise (i) If = max, 0, find the mean and variance of ; (ii) If = 2 + 3, + 4, find. Proof. On using Theorem 3.2 (ii) we get, for > 0, = max, 0 27

28 = max, 0 = 9 = It follows that = 1, = 9 4 and Var = = 5 4 (iii) We have = = 2 = and, =, = = 18 Therefore, = 2 + 3, + 4 = 2 + 3, + 4 Example 3.6 = 28

29 Let be random variable with p.m.f. where 1, 2,,, 0, 1 and = 1. =, if 0, 1,,, 0, otherwise (i) (ii) (iii) For 1, 2,, find, where = is called the -th factorial moment of, = 1,2, ; Find mean and variance of ; Let = Find. Solution. (i) Fix 1, 2,,. Then = = 1 + 1! = 1 + 1!! = = = = (ii) Using (i), we get = = 1 = = 1 Therefore, = 1 + =

30 and Var = =. (iii) For R, we have = = = +. Therefore, = = = We are familiar with the Laplace transform of a given real-valued function defined on R. We also know that, under certain conditions, the Laplace transform of a function determines the function almost uniquely. In probability theory the Laplace transform of a p.d.f./p.m.f. of a random variable plays an important role and is referred to as moment generating function (of probability distribution) of random variable. Definition 3.3 Let be a random variable and let = R: = is finite. Define : R by =,. (i) (ii) We call the function the moment generating function (m.g.f.) (of probability distribution) of random variable ; We say that the m.g.f. of a random variable exists if there exists a positive real number such that, (i.e., if = is finite in an interval containing 0). Note that 0 = 1 and, therefore, = R: is finite. Moreover, using Theorem 3.3 (ii)-(iii), we have > 0,. Also if = exists and is finite on an interval,, > 0, then for any real constants and the m.g.f. of = + also 30

31 exists and = = = =,,, with the convention that 0 = The name moment generating function to the transform is derived from the fact that can be used to generate moments of random variable, as illustrated in the following theorem. Theorem 3.4 Let be a random variable with m.g.f. that is finite on an interval,, for some > 0 (i.e., m.g.f. of exists). Then, (i) for each 1, 2,, = is finite; (ii) for each 1, 2, exists on,, and for each 1, 2,, = (iii) = 0, where 0 =, the r-th derivative of at the point 0; =!,,. Proof. We will provide the proof for the case where is of absolutely continuous type. The proof for the case of discrete type follows in the similar fashion with integral signs being replaced by summation sings. (i) We have <,,. Therefore, <,, and <,, <,, and <,, <,, and <,, <,, and <,, 31

32 ) <,, ), i.e., <,, ). Fix 1, 2, and, ) 0. Then lim = 0 and therefore there exists a positive real number, such that <, whenever >,. Thus we have ) = ) = ) + ),,, ) + ),,, <. + ) (ii) Fix 1, 2,. Then, for, ), and ) = ) ) ) = ). Under the assumption that ) = ) <,, ), using arguments of advanced calculus, it can be shown that the derivatives exists and can be passed through the integral sign. Therefore ) ) = ) = ) 32

33 0 = =. (iii) Fix 1, 2,. Then, for,, = =!. Under the assumption that = <,,, using arguments of advanced calculus, it can be shown that the integral sign can be passed through the summation sign, i.e., Corollary 3.2 =! =!. Under the notation and assumptions of Theorem 3.4, define :, R by = ln,,. Then = 0 and = Var = 0, where denotes the r-th 1, 2 derivative of. Proof. We have, for,, = and =. Using the facts that 0 = 1 and 0 =, 1, 2, we get 0 = 0 0 =, 33

34 and 0 = = = Var. Example 3.7 Let be a random variable with p.m.f. where > 0. =, if 0, 1, 2,!, 0, otherwise (i) (ii) (iii) Find the m.g.f., = R: <, of. Show that possesses moment of all orders. Find the mean and variance of ; Find = ln,. Hence find the mean and variance of ; What are the first four terms in the power series expansion of around the point 0. Solution. (i) We have = = = = =, R.!! Since = R: < = R, by Theorem 3.4 (i), for every 1, 2,, = is finite. Clearly Therefore, = and = 1 +, R. = 0 =, = 0 = 1 +, 34

35 and Var = =. (ii) We have, for R, = ln = 1, = =. Therefore, = 0 = and Var = 0 =. (iii) We have = , R = = 0 = Since = R: < = R, by Theorem 3.4 (iii), we have Example 3.8 = ! + 3! + Let be a random variable with p.d.f. = ! , R. 3! =, if > 0 0, otherwise. (i) (ii) (iii) Find the m.g.f., = R: < of. Show that posseses moments of all orders. Find the mean and variance of ; Find = ln,. Hence find the mean and variance of ; Expand as a power series around the point 0 and hence find, 1, 2,. Solution. (i) We have = = = <, if < 1. Clearly = R: < =, 1 1, 1and = 1, < 1. By Theorem 3.4 (i), for every 1, 2,, is finite. Clearly 35

36 = 1 and = 21, < 1, = 0 = 1, = 0 = 2, and Var = = 1. (ii) We have = ln = ln1, < 1 = 1 1 and 1 =, < 1 1 = 0 = 1 and Var = 0 = 1. (iii) We have = 1 =, 1, 1. Example 3.9 Since = R: < =, 1 1, 1, using Theorem 3.4 (iii), we conclude that = coefficient of =!. Let be a random variable with p.d.f. Show that the m.g.f. of does not exist.! in the power series expansion of around 0 = 1 π. 1, < <. 1 + Solution. From Example 3.4 we know that the expected value of is not finite. Therefore, using Theorem 3.4 (i), we conclude that the m.g.f. of does not exist. 4. Properties of Random Variables Having the Same Distribution We begin this section with the following definition. 36

37 Definition 4.1 Two random variables and, defined on the same probability space Ω, F,, are said to have the same distribution (written as = ) if they have the same distribution function, i.e., if =, R. Theorem 4.1 (i) (ii) Let and be random variables of discrete type with p.m.f.s and respectively. Then = if, and only if, =, R; Let and be random variables of absolutely continuous type. Then = if, and only if, there exist versions of p.d.f.s and of and, respectively, such that =, R. Proof. (i) Suppose that =, R. Then, clearly, =, R, and therefore =. Conversely suppose that =, i.e., = =, say, R. Then R: > 0 = R: > 0 = R: > 0 = =, say. Moreover,, if = = 0, otherwise. (ii) Suppose that =, R, for some versions of p.d.f.s of and of and respectively. Then, clearly, =, R and therefore =. Conversely suppose that =, i.e., suppose that = =, say, R. For simplicity suppose that is differentiable everywhere except possibly on a countable set and = 1. Using Remark 4.2 (vii), Module 2, it follows that both and are of absolutely continuous type with a common (version of) p.d.f. =, if 0, otherwise 37

38 As a consequence of the above theorem we have the following corollary. Theorem 4.2 Let and be two random variables, of either discrete type or of absolutely continuous type, with =. Then, (i) (ii) for any Borel function h, h = h, provided the expectations are finite; for any Borel function, =. Proof. (i) Since =, we have = =, say, R Case I. is of discrete type. Since =, using Theorem 4.1 (i), it follows that = =, say, and = =, say, R. Therefore, Case II. is of absolutely continuous type. h = h = h = h. For simplicity assume that is differentiable everywhere except possibly on a finite set. Using Remark 4.2 (vii), Module 2, we may take Therefore, = =, if 0, if h = h = h 38

39 (ii) Fix R. On taking in (i), we get = h. h =, = 1, if 0, if >, =, =, R =. Example 4.1 (i) Let be a random variable with p.m.f. = 1 2, if 0, 1,,, 0, otherwise where is a given positive integer. Let =. Show that = and hence show that =. (ii) Let be a random variable with p.d.f. Solution. = 2, < <, and let =. Show that = and hence show that = 0, 0, 1,. (i) Clearly is finite. Using Example 2.3 it follows that the p.m.f. of = is given by = = = 1 2, if 0, 1,, 0, otherwise =, R, 39

40 i.e., =. Hence, using Theorem 4.2 (i), = = = = 2. (ii) Using Corollary 2.2, it can be shown that = is a random variable of absolutely continuous type with p.d.f. = 2 =, < <. It follows that =. It can be shown that is finite for every > 1. Therefore Definition 4.2 = = = 0. A random variable is said to have a symmetric distribution about a point R if =. Theorem 4.3 Let be a random variable having p.d.f./p.m.f. and distribution function. Let R. (i) Then the distribution of is symmetric about if, and only if, = +, R ; (ii) Then the distribution of is symmetric about if, and only if, + + = 1, R i. e., if and only if, + = ; (iii) Then the distribution of is symmetric about if, and only if, the distribution of = is symmetric about 0; (iv) If the distribution of is symmetric about, then ; (v) If the distribution of is symmetric about and the expected value of is finite, then = ; (vi) If the distribution of is symmetric about 0, then = 0, 0, 1, 2,, provided the involved expectations exist. Proof. For simplicity we will assume that if is of absolutely continuous type then its distribution function is differentiable everywhere except possibly on a countable set. 40

41 (i) Let = and =. Then the p.d.f.s/p.m.f.s of and are given by = +, R and =, R, respectively. Now, under the hypothesis, distribution of is symmetric about = =, R + =, R. (ii) Let = and = so that the distribution functions of and are given by = +, R, and = 1, R. Therefore, under the hypothesis, = =, R + + = 1, R. (iii) Clearly, distribution of is symmetric about = = =. (iv) Using (ii), we have distribution of is symmetric about + + = 1, R + = 1 (v) since. 1 2, Suppose that = and <. Then = and therefore =. (vi) Suppose that =. Then =, 0,1,2,, provided the expectation exist. Therefore, =, 0, 1, 2, = 0, 0, 1, 2,. Theorem 4.4 Let and be random variables having m.g.f.s and respectively. Suppose that there exists a positive real number such that =,,. Then =. 41

42 Proof. We will provide the proof for the special case where and are of discrete type and = 0, 1, 2,, as the proof for general and is involved. We have =,, = = =,, = = =,,. We know that if two power series or polynomials match over an interval then they have the same coefficients. It follows that = = =, 0,1,2,, i. e., and have the same p.m.f.. Now the result follows on using Theorem 4.1 (i). Example 4.2 Let R and > 0 be real constants and let, be a random variable having a p.d.f., = (i) Show that, is a p.d.f.; 1 σ 2 e, < < 4.1 (ii) Show that the probability distribution function of, is symmetric about. Hence find, ; (iii) Find the m.g.f. of, and hence find the mean and variance of, ; (iv) Let, =, +, where 0 and are real constants. Using the m.g.f. of,, show Solution. that the p.d.f. of, is, = 1 a σ 2 e, < <. (i) Clearly, 0, R. Also,, = e = 1 2 e on making the transformation μ = Clearly 0 and =, say. 42

43 = 1 2 e = 1 2π 1 2. On making the transformation = cos, = sin, > 0, 0 < 2 (so that the Jacobian of the transformation is r) we get = 1 2 = = = 1. e Since 0, it follows that = 1 and thus, is a p.d.f.. (ii) Clearly,, μ =, μ + = 1 σ 2 e, R. Using Theorem 4.3 (i) and (v) it follows that the distribution of, is symmetric about μ and E, = μ. (iii) For R, =, = e 1 σ 2 e d = 1 2 e e d = = e, e dt 43

44 since, by (i), e d = σ 2π, R and > 0. Thus, for R, ψ, t = ln M, t = μt + σ t EX =ψ, 0 = μ and VarX =ψ, 0 = σ. (iv) From the discussion following Definition 3.3 we have, for R, M, t =, =, = =, 2 Therefore the p.d.f. of, is given by, =,., =, = 1 2 R. In the statistical literature, the probability distribution of the random variable, having a p.d.f.,, defined by (4.1), is called the normal distribution (or Gaussian distribution) with mean and variance (denoted by,, ). Various properties of this distribution are further discussed in Module 5. Example 4.3 Let 0,1 and let be a random variable with p.m.f. =, if 0, 1,,, 4.2 0, otherwise where is a given positive integer and = 1. 44

45 (i) Find the m.g.f. of and hence find the mean and variance of, 0,1; (ii) Let =, p 0,1. Using the m.g.f. of show that the p.m.f. of is Solution. = 1, if 0, 1,,. 0, otherwise (i) From the solution of Example 3.6 (iii), it is clear that the m.g.f. of is given by = 1 +, R. Therefore, for t R, = ln = ln1 +, R, =, R, 1 + = , R = 0 = and Var = 0 = 1. (ii) For t R = = = 1 + = + 1 =, i.e., =, 0,1. Therefore the p.m.f. of is given by = = 1, if 0, 1,,. 0, otherwise The probability distribution of the random variable having p.m.f., defined by (4.2), is called a binomial distribution with trials and success probability (denoted by 45

46 Bin, ). The binomial distribution and other related distributions are discussed in more detail in Module Probability and Moment inequalities Let be a random variable defined on a probability space Ω, F,, and let B be a Borel set. In many situation cannot be explicitly evaluated and therefore some estimate of this probability may be desired. For example if a random variable has the p.d.f. = 1 2, < <, then > 2 = (5.1) cannot be explicitly evaluated and, therefore, an estimate of this probability may be desired. To estimate this probability one has to either resort to numerical integration or use some other estimation procedure. Inequalities are popular estimation procedures and they play an important role in probability theory. Theorem 5.1 Let be a random variable and let : [0, R be a non-negative and non-decreasing function such that the expected value of is finite. Then, for any > 0 for which > 0,. Proof. We will provide the proof for the case when is of absolutely continuous type. The proof for the discrete case follows in the similar fashion with integral signs replaced by summation signs. Fix > 0 such that > 0. Define = R, so that, for,. Then = since, R 46

47 , R, as = = Ε. Corollary 5.1 Let be a random variable. (i) (ii) Proof. (i) (ii) (Markov Inequality) Suppose that <, for some > 0. Then, for any > 0, (Chebyshev Inequality) Suppose that has finite first two moments. If = and = Var 0. Then for any > 0,. Fix > 0 and > 0 and let =, 0. Clearly is a non-negative and nondecreasing function. Using Theorem 5.1, we get. Using (i) on =, for = 2, we get Example: 5.1 =. Let us revisit the problem of estimating > 2, defined by (5.1). Using Example 4.2 (iii), we have = = 0 and = Var = 1. Moreover, using Example 4.2 (ii), =. Consequently > 2 = > 2 = < 2, i.e., > 2 = > 2 2 = 2. 2 Now, using the Chebyshev inequality we have 47

48 > 2 = = 1 8 = The exact value of > 2, obtained using numerical integration, is The following example illustrates that bounds provided in Theorem 5.1 and Corollary 5.1 are tight, i.e., the upper bounds provided there in may be attained. Example 5.2 Let be a random variable with p.m.f. 1 8, if 1,1 = 3., if = 0 4 0, otherwise Clearly = and, therefore, using the Markov inequality we have The exact probability is 1 = 1 4. Definition = 1,1 = 1 4. A random variable is said to be degenerate at a point R if = = 1. Suppose that a random variable is degenerate at R. Then clearly is of discrete type with support =, distribution function. and p.m.f. = 0, < 1,, = 1, = 0, otherwise. Note that a random variable is degenerate at a point R if, and only if, = and Var = 0. 48

49 Theorem 5.2 (Jensen Inequality) Let R be an interval and let : R be a twice differentiable function such that its second order derivative is continuous on and 0, R. Let be a random variable with support and finite expectation. Then. If > 0,, then the inequality above is strict unless is a degenerate random variable. Proof. Let =. On expanding into a Taylor series about we get = + + ",, 2! for some between and. Since " 0,, it follows that +, 5.2 +, + = =. Clearly the inequality in (5.2) is strict unless = 0, i. e., = = 1 (using Corollary 3.1 (ii)). Example 5.3 Let be a random variable with support, an interval in R. Then (i) taking =, R, in Theorem 5.2; (ii) ln ln,provided 0, taking = ln, R, in Theorem 5.2; (iii) taking =, R, in Theorem 5.2; (iv), if 0, taking =, R in Theorem 5.2 ; provided the involved expectations exist. Definition

50 Let be a random variable with support 0,. Then, provided they are finite, is called the arithmetic mean of, is called the geometric mean of, and is called harmonic mean ( of. Example 5.4 (i) (-- inequality) Let be a random variable with support 0,. Then provided the expectations are finite. 1, (ii) Let,, be positive real constants and let,, be another set of positive real constants such that = 1. Then 1 Solution. (i) From Example 5.3 (ii) we have ln ln (5.3) Using (5.3) on,we get 1 = The assertion now follows on combined (5.3) and (5.4).. (5.4) (ii) Let be a discrete type random variable with support =, and = =, = 1,,. Clearly, = > 0 an = = = 1. On using (i), we get 50

51 Descriptive Measures of Probability Distributions Let be a random variable defined on a probability space Ω, F,, associated with a random experiment E. Let and denote, respectively, the distribution function and the p.d.f./p.m.f. of. The probability distribution (i.e., the distribution function/p.d.f./p.m.f.) of describes the manner in which the random variable takes values in different Borel sets. It may be desirable to have a set of numerical measures that provide a summary of the prominent features of the probability distribution of. We call these measures as descriptive measures. Four prominently used descriptive measures are: (i) Measures of central tendency (or location), also referred to as averages; (ii) measures of dispersion; (iii) measures of skewness, and (iv) measures of kurtosis. I. Measures of Central Tendency A measure of central tendency or location (also called an average) gives us the idea about the central value of the probability distribution around which values of the random variable are clustered. Three commonly used measures of central tendency are mean, median and mode. I (a). Mean. Recall (Definition 3.2 (i)) that the mean (of probability distribution) of a random variable is given by =. We have seen that the mean of a probability distribution gives us idea about the average observed value of in the long run (i.e., the average of observed values of when the random experiment is repeated a large number of times). Mean seems to be the best suited average if the distribution is symmetric about a point (i.e., =, in which case = provided it is finite), values in the neighborhood of occur with high probabilities, and as we move away from in either direction decreases. 51

52 Because of its simplicity mean is the most commonly used average (especially for symmetric or nearly symmetric distributions). Some of the demerits of this measure are that in some situations this may not be defined (Examples 3.2 and 3.4) and that it is very sensitive to presence of a few extreme values of which are different from other values of (even though they may occur with small positive probabilities). So this measure should be used with caution if probability distribution assigns positive probabilities to a few Borel sets having some extreme values. I (b). Median. A real number satisfying 1 2, i. e., < 1, 2 is called the median (of the probability distribution) of. Clearly if is the median of a probability distribution then, in the long run (i. e., when the random experiment E is repeated a large number of times), the values of on either side of in are observed with the same frequency. Thus the median of a probability distribution, in some sense, divides into two equal parts each having the same probability of occurrence. It is evident that if is of continuous type then the median m is given by = 1/2 and it may not be unique. Also, for some distributions (especially for distributions of discrete type random variable) it may happen that < and R: = 1/2 = [,, for some < < <, so that the median is again not unique. In such situations we define the median to be = inf R: 1/2. For random variables having a symmetric probability distribution it is easy to verify that the mean and the median coincide (see Problem 33). Unlike the mean, the median of a probability distribution is always defined. Moreover the median is not affected by a few extreme values as it takes into account only the probabilities with which different values occur and not their numerical values. As a measure of central tendency the median is preferred over the mean if the distribution is asymmetric and a few extreme observations are assigned positive probabilities. However the fact that the median does not at all take into account the numerical values of is one of its demerits. Another disadvantage with median is that for many probability distributions it is not easy to evaluate (especially for distributions whose distribution functions do not have closed forms). I (c). Mode. Roughly speaking the mode of a probability distribution is the value that occurs with highest probability and is defined by = sup :. Clearly if is the mode of a probability distribution of then, in the long run, either or a value in the neighborhood of is observed with maximum frequency. Mode is easy to understand and easy to calculate. Normally, it can be found by just inspection. Note that a probability distribution may have more than one mode which may be far apart. Moreover mode does not take into account the numerical values of and it also does not take into account the 52

53 probabilities associated with all the values of. These are crucial deficiencies of mode which make it less preferable over mean and median. A probability distribution with one (two / three) mode(s) is called an unimodal (bimodal/trimodal) distribution. A distribution with multiple modes is called a multimodal distribution. II. Measures of Dispersion Measures of central tendency give us the idea about the location of only central part of the distribution. Other measures are often needed to describe a probability distribution. The values assumed by a random variable usually differ from each other. The usefulness of mean or median as an average is very much dependent on the variability (or dispersion) of values of around mean or median. A probability distribution (or the corresponding random variable is said to have a high dispersion if its support contains many values that are significantly higher or lower than the mean or median value. Some of the commonly used measures of dispersion are standard deviation, quartile deviation (or semi-inter-quartile range) and coefficient of variation. II (a). Standard Deviation. Recall (Definition 3.2 (v)) that the variance (of the probability distribution) of a random variable is defined by =, where = is the mean (of the probability distribution) of. The standard deviation (of the probability distribution) of is defined by = =. Clearly the variance and the standard deviation give us the idea about the average spread of values of around the mean. However, unlike the variance, the unit of measurement of standard deviation is the same as that of. Because of its simplicity and intuitive appeal, standard deviation is the most widely used measure of dispersion. Some of the demerits of standard deviation are that in many situations it may not be defined (distribution for which second moment is not finite) and that it is sensitive to presence of a few extreme values of which are different from other values. A justification for having the mean in place of median or any other average in the definition of = is that R (Problem 32). II (b). Mean Deviation. Let be a suitable average. The mean deviation (of probability distribution) of around average is defined by MD =. Among various mean deviations, the mean deviation about the median is more preferable than the others. A reason for this preference is the fact that for any random variable,md = =MD, R (Problem 24). Since a natural distance between and is, as a measure of dispersion, the mean deviation about median seems to be more appealing than the standard deviation. Although the mean deviation about median (or mean) has more intuitive appeal than the standard deviation, in most situations, it is not easy to evaluate. Some of the other demerits of mean deviations are that in many situations they may 53

54 not be defined and that they are sensitive to presence of a few extreme values of which are different from other values. II (c). Quartile Deviation. A common drawback with the standard deviation and mean deviations, as measures of dispersion, is that they are sensitive to presence of a few extreme values of. Quartile deviation measures the spread in the middle half of the distribution and is therefore not influenced by extreme values. Let and be real numbers such that 1 4 and 3 4, i. e., < and <. The quantities and are called, respectively, the lower and upper quartiles of the probability distribution of random variable. Clearly if, and are respectively the lower quartile, the median and the upper quartile of a probability distribution then they divide the probability distribution in four parts so that, in the long run (i.e., when the random experiment E is repeated a large number of times), twenty five percent of the observed values of are expected to be less than, fifty percent of the observed values of are expected to be less than and seventy five percent of the observed values of are expected to be less than. The quantity IQR = is called the inter quartile range of the probability distribution of and the quantity QD = is called the quartile deviation or the semi-interquartile range of the probability distribution of. It can be seen that if is of absolutely continuous type then and are given by = and = and they may not be unique. Also, for some distributions (especially for distributions of discrete type random variables) it may happen that < and/or R: = 1/4 = [, ( < and/or R: = 3/4 = [, ) for some < < < (for some < < <, so that ( ) is again not uniquely defined. In such situations we define = inf R: 1/4 ( = inf R: 3/4 ). For random variables having a symmetric probability distribution it is easy to verify that = + /2 (Problem 33). Although, unlike the standard deviation and the mean deviation, quartile deviation is not sensitive to presence of some extreme values of a major drawback with the quartile deviation is that it ignores the tails of the probability distribution (which constitute 50% of the probability distribution). Note that the quartile deviation depends on the units of measurement of random variable and thus it may not be an appropriate measure for comparing dispersions of two different probability distributions. For comparing dispersions of two different probability distributions a normalized measure such as 54

55 CQD = = + seems to be more appropriate. The quantity CQD is called the coefficient of quartile deviation of the probability distribution of. Clearly the coefficient of quartile deviation is independent of units of measurement and thus it can be used to be compare dispersion of two different probability distributions. II (d). Coefficient of Variation. Like quartile deviation, standard deviation also depends on the units of measurement of random variable and thus it is not an appropriate measure for comparing dispersions of two different probability distributions. Let and, respectively, be the mean and the standard deviation of the distribution of. Suppose that 0. The coefficient of variation of the probability distribution of is defined by =. Clearly the coefficient of variation measures the variation per unit of mean and is independent of units. Therefore it seems to be an appropriate measure to compare dispersions of two different probability distributions. A disadvantage with the coefficient of variation is that when the mean is close to zero it is very sensitive to small changes in the mean. III. Measures of Skewness of a probability distribution is a measure of asymmetry (or lack of symmetry). Recall that the probability distribution of random variable is said to be symmetric about point if =. In that case = (provided it exists) and + =, R. Evidently, for symmetric distributions, the shape of the p.d.f./p.m.f. on the left of is the mirror image of that on the right side of. It can be shown that, for symmetric distributions the mean and the media coincide (Problem 33). We say that a probability distribution is positively skewed if the tail on the right side of the p.d.f./p.m.f. is longer than that on the left side of the p.d.f./p.m.f. and bulk of the values lie on the left side of the mean. Clearly a positively skewed distribution indicates presence of a few high values of which pull up the value of the mean resulting in mean larger than the median and the mode. For unimodal and positively skewed distribution we normally have Mode <Median <Mean. Similarly we say that a probability distribution is negatively skewed if the tail on the left side of the p.d.f./p.m.f. is longer than that on the right side of the p.d.f./p.m.f. and bulk of the values lie on the right side of the mean. Clearly a negatively skewed distribution indicates presence of a few low values of which pull down the value of the mean resulting in mean smaller than the median and the mode. For unimodal and negatively skewed distribution we normally have Mean< Median <Mode. Let 55

56 and, respectively, be the mean and the standard deviation of and let = / be the standardized variable (independent of units of measurement). A measure of skewness of the probability distribution of is defined by = = =. The quantity is simply called the coefficient of skewness. Clearly, for symmetric distributions, = 0 (Theorem 4.3 (vi)). However the converse may not be true, i.e., there are examples of skewed probability distributions for which = 0. A large positive value of indicates that the data is positively skewed and a small negative value of indicates that the data is negatively skewed. A measure of skewness can also be based on quartiles. Let,, and denote respectively the lower quartile, the median, the upper quartile and the mean of the probability distribution of. We know that for random variables having a symmetric probability = =,.., =. For positively (negatively) skewed distribution we will have > <. Thus one may also define a measure of skewness based on = 2 +. To make this quantity independent of units of measurement one may consider = 2 + as a measure of skewness. The quantity is called the Yule coefficient of skewness. IV. Measures of Kurtosis For real constants R and > 0, let, be a random variable having the p.d.f., = 1 2, < <, < <, > 0, i.e.,,, (see Example 4.2 and discussion following it). We have seen (Example 4.2 (iii)) that and are respectively the mean and the variance of the distribution of,. We call the probability distribution corresponding to p.d.f., as the normal distribution with mean and variance (denoted by,. We know that, distribution is symmetric about (cf. Example 4.2 (ii)). Also it is easy to verify that, distribution is unimodal with as the common value of mean, median and mode. Kurtosis of the probability distribution of is a measure of peakedness and thickness of tail of p.d.f. of relative to the peakedness and thickness of tails of the p.d.f. of normal distribution. A distribution is said to have higher (lower) kurtosis than the normal distribution if its p.d.f., in comparison with the p.d.f. of a normal distribution, has a sharper (rounded) peak and longer, fatter (shorter, thinner) tails. Let and 56

Notes 6 : First and second moment methods

Notes 6 : First and second moment methods Notes 6 : First and second moment methods Math 733-734: Theory of Probability Lecturer: Sebastien Roch References: [Roc, Sections 2.1-2.3]. Recall: THM 6.1 (Markov s inequality) Let X be a non-negative

More information

MODULE 3 FUNCTION OF A RANDOM VARIABLE AND ITS DISTRIBUTION LECTURES PROBABILITY DISTRIBUTION OF A FUNCTION OF A RANDOM VARIABLE

MODULE 3 FUNCTION OF A RANDOM VARIABLE AND ITS DISTRIBUTION LECTURES PROBABILITY DISTRIBUTION OF A FUNCTION OF A RANDOM VARIABLE Topics MODULE 3 FUNCTION OF A RANDOM VARIABLE AND ITS DISTRIBUTION LECTURES 2-6 3. FUNCTION OF A RANDOM VARIABLE 3.2 PROBABILITY DISTRIBUTION OF A FUNCTION OF A RANDOM VARIABLE 3.3 EXPECTATION AND MOMENTS

More information

3 Integration and Expectation

3 Integration and Expectation 3 Integration and Expectation 3.1 Construction of the Lebesgue Integral Let (, F, µ) be a measure space (not necessarily a probability space). Our objective will be to define the Lebesgue integral R fdµ

More information

The Canonical Gaussian Measure on R

The Canonical Gaussian Measure on R The Canonical Gaussian Measure on R 1. Introduction The main goal of this course is to study Gaussian measures. The simplest example of a Gaussian measure is the canonical Gaussian measure P on R where

More information

IAM 530 ELEMENTS OF PROBABILITY AND STATISTICS LECTURE 3-RANDOM VARIABLES

IAM 530 ELEMENTS OF PROBABILITY AND STATISTICS LECTURE 3-RANDOM VARIABLES IAM 530 ELEMENTS OF PROBABILITY AND STATISTICS LECTURE 3-RANDOM VARIABLES VARIABLE Studying the behavior of random variables, and more importantly functions of random variables is essential for both the

More information

Module 1. Probability

Module 1. Probability Module 1 Probability 1. Introduction In our daily life we come across many processes whose nature cannot be predicted in advance. Such processes are referred to as random processes. The only way to derive

More information

Chapter 6 Expectation and Conditional Expectation. Lectures Definition 6.1. Two random variables defined on a probability space are said to be

Chapter 6 Expectation and Conditional Expectation. Lectures Definition 6.1. Two random variables defined on a probability space are said to be Chapter 6 Expectation and Conditional Expectation Lectures 24-30 In this chapter, we introduce expected value or the mean of a random variable. First we define expectation for discrete random variables

More information

2. Variance and Higher Moments

2. Variance and Higher Moments 1 of 16 7/16/2009 5:45 AM Virtual Laboratories > 4. Expected Value > 1 2 3 4 5 6 2. Variance and Higher Moments Recall that by taking the expected value of various transformations of a random variable,

More information

Chapter 5 Random vectors, Joint distributions. Lectures 18-23

Chapter 5 Random vectors, Joint distributions. Lectures 18-23 Chapter 5 Random vectors, Joint distributions Lectures 18-23 In many real life problems, one often encounter multiple random objects. For example, if one is interested in the future price of two different

More information

Notes on Mathematical Expectations and Classes of Distributions Introduction to Econometric Theory Econ. 770

Notes on Mathematical Expectations and Classes of Distributions Introduction to Econometric Theory Econ. 770 Notes on Mathematical Expectations and Classes of Distributions Introduction to Econometric Theory Econ. 77 Jonathan B. Hill Dept. of Economics University of North Carolina - Chapel Hill October 4, 2 MATHEMATICAL

More information

Summarizing Measured Data

Summarizing Measured Data Performance Evaluation: Summarizing Measured Data Hongwei Zhang http://www.cs.wayne.edu/~hzhang The object of statistics is to discover methods of condensing information concerning large groups of allied

More information

Estimates for probabilities of independent events and infinite series

Estimates for probabilities of independent events and infinite series Estimates for probabilities of independent events and infinite series Jürgen Grahl and Shahar evo September 9, 06 arxiv:609.0894v [math.pr] 8 Sep 06 Abstract This paper deals with finite or infinite sequences

More information

STAT 7032 Probability Spring Wlodek Bryc

STAT 7032 Probability Spring Wlodek Bryc STAT 7032 Probability Spring 2018 Wlodek Bryc Created: Friday, Jan 2, 2014 Revised for Spring 2018 Printed: January 9, 2018 File: Grad-Prob-2018.TEX Department of Mathematical Sciences, University of Cincinnati,

More information

Math212a1413 The Lebesgue integral.

Math212a1413 The Lebesgue integral. Math212a1413 The Lebesgue integral. October 28, 2014 Simple functions. In what follows, (X, F, m) is a space with a σ-field of sets, and m a measure on F. The purpose of today s lecture is to develop the

More information

Chapter 6. Convergence. Probability Theory. Four different convergence concepts. Four different convergence concepts. Convergence in probability

Chapter 6. Convergence. Probability Theory. Four different convergence concepts. Four different convergence concepts. Convergence in probability Probability Theory Chapter 6 Convergence Four different convergence concepts Let X 1, X 2, be a sequence of (usually dependent) random variables Definition 1.1. X n converges almost surely (a.s.), or with

More information

Chapter 4. Displaying and Summarizing. Quantitative Data

Chapter 4. Displaying and Summarizing. Quantitative Data STAT 141 Introduction to Statistics Chapter 4 Displaying and Summarizing Quantitative Data Bin Zou (bzou@ualberta.ca) STAT 141 University of Alberta Winter 2015 1 / 31 4.1 Histograms 1 We divide the range

More information

Some Background Material

Some Background Material Chapter 1 Some Background Material In the first chapter, we present a quick review of elementary - but important - material as a way of dipping our toes in the water. This chapter also introduces important

More information

STAT/MATH 395 A - PROBABILITY II UW Winter Quarter Moment functions. x r p X (x) (1) E[X r ] = x r f X (x) dx (2) (x E[X]) r p X (x) (3)

STAT/MATH 395 A - PROBABILITY II UW Winter Quarter Moment functions. x r p X (x) (1) E[X r ] = x r f X (x) dx (2) (x E[X]) r p X (x) (3) STAT/MATH 395 A - PROBABILITY II UW Winter Quarter 07 Néhémy Lim Moment functions Moments of a random variable Definition.. Let X be a rrv on probability space (Ω, A, P). For a given r N, E[X r ], if it

More information

Lecture 9. = 1+z + 2! + z3. 1 = 0, it follows that the radius of convergence of (1) is.

Lecture 9. = 1+z + 2! + z3. 1 = 0, it follows that the radius of convergence of (1) is. The Exponential Function Lecture 9 The exponential function 1 plays a central role in analysis, more so in the case of complex analysis and is going to be our first example using the power series method.

More information

STAT 3610: Review of Probability Distributions

STAT 3610: Review of Probability Distributions STAT 3610: Review of Probability Distributions Mark Carpenter Professor of Statistics Department of Mathematics and Statistics August 25, 2015 Support of a Random Variable Definition The support of a random

More information

Week 2. Review of Probability, Random Variables and Univariate Distributions

Week 2. Review of Probability, Random Variables and Univariate Distributions Week 2 Review of Probability, Random Variables and Univariate Distributions Probability Probability Probability Motivation What use is Probability Theory? Probability models Basis for statistical inference

More information

A LITTLE REAL ANALYSIS AND TOPOLOGY

A LITTLE REAL ANALYSIS AND TOPOLOGY A LITTLE REAL ANALYSIS AND TOPOLOGY 1. NOTATION Before we begin some notational definitions are useful. (1) Z = {, 3, 2, 1, 0, 1, 2, 3, }is the set of integers. (2) Q = { a b : aεz, bεz {0}} is the set

More information

Sample Spaces, Random Variables

Sample Spaces, Random Variables Sample Spaces, Random Variables Moulinath Banerjee University of Michigan August 3, 22 Probabilities In talking about probabilities, the fundamental object is Ω, the sample space. (elements) in Ω are denoted

More information

1 Variance of a Random Variable

1 Variance of a Random Variable Indian Institute of Technology Bombay Department of Electrical Engineering Handout 14 EE 325 Probability and Random Processes Lecture Notes 9 August 28, 2014 1 Variance of a Random Variable The expectation

More information

1 Directional Derivatives and Differentiability

1 Directional Derivatives and Differentiability Wednesday, January 18, 2012 1 Directional Derivatives and Differentiability Let E R N, let f : E R and let x 0 E. Given a direction v R N, let L be the line through x 0 in the direction v, that is, L :=

More information

Chapter 4. Chapter 4 sections

Chapter 4. Chapter 4 sections Chapter 4 sections 4.1 Expectation 4.2 Properties of Expectations 4.3 Variance 4.4 Moments 4.5 The Mean and the Median 4.6 Covariance and Correlation 4.7 Conditional Expectation SKIP: 4.8 Utility Expectation

More information

Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension. n=1

Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension. n=1 Chapter 2 Probability measures 1. Existence Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension to the generated σ-field Proof of Theorem 2.1. Let F 0 be

More information

NATIONAL UNIVERSITY OF SINGAPORE Department of Mathematics MA4247 Complex Analysis II Lecture Notes Part II

NATIONAL UNIVERSITY OF SINGAPORE Department of Mathematics MA4247 Complex Analysis II Lecture Notes Part II NATIONAL UNIVERSITY OF SINGAPORE Department of Mathematics MA4247 Complex Analysis II Lecture Notes Part II Chapter 2 Further properties of analytic functions 21 Local/Global behavior of analytic functions;

More information

If g is also continuous and strictly increasing on J, we may apply the strictly increasing inverse function g 1 to this inequality to get

If g is also continuous and strictly increasing on J, we may apply the strictly increasing inverse function g 1 to this inequality to get 18:2 1/24/2 TOPIC. Inequalities; measures of spread. This lecture explores the implications of Jensen s inequality for g-means in general, and for harmonic, geometric, arithmetic, and related means in

More information

Random variables. DS GA 1002 Probability and Statistics for Data Science.

Random variables. DS GA 1002 Probability and Statistics for Data Science. Random variables DS GA 1002 Probability and Statistics for Data Science http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall17 Carlos Fernandez-Granda Motivation Random variables model numerical quantities

More information

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015 Part IA Probability Theorems Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.

More information

Section 21. The Metric Topology (Continued)

Section 21. The Metric Topology (Continued) 21. The Metric Topology (cont.) 1 Section 21. The Metric Topology (Continued) Note. In this section we give a number of results for metric spaces which are familar from calculus and real analysis. We also

More information

17. Convergence of Random Variables

17. Convergence of Random Variables 7. Convergence of Random Variables In elementary mathematics courses (such as Calculus) one speaks of the convergence of functions: f n : R R, then lim f n = f if lim f n (x) = f(x) for all x in R. This

More information

Sequences of Real Numbers

Sequences of Real Numbers Chapter 8 Sequences of Real Numbers In this chapter, we assume the existence of the ordered field of real numbers, though we do not yet discuss or use the completeness of the real numbers. In the next

More information

Integration on Measure Spaces

Integration on Measure Spaces Chapter 3 Integration on Measure Spaces In this chapter we introduce the general notion of a measure on a space X, define the class of measurable functions, and define the integral, first on a class of

More information

Introduction to Probability

Introduction to Probability LECTURE NOTES Course 6.041-6.431 M.I.T. FALL 2000 Introduction to Probability Dimitri P. Bertsekas and John N. Tsitsiklis Professors of Electrical Engineering and Computer Science Massachusetts Institute

More information

U e = E (U\E) e E e + U\E e. (1.6)

U e = E (U\E) e E e + U\E e. (1.6) 12 1 Lebesgue Measure 1.2 Lebesgue Measure In Section 1.1 we defined the exterior Lebesgue measure of every subset of R d. Unfortunately, a major disadvantage of exterior measure is that it does not satisfy

More information

Analysis Finite and Infinite Sets The Real Numbers The Cantor Set

Analysis Finite and Infinite Sets The Real Numbers The Cantor Set Analysis Finite and Infinite Sets Definition. An initial segment is {n N n n 0 }. Definition. A finite set can be put into one-to-one correspondence with an initial segment. The empty set is also considered

More information

18.175: Lecture 3 Integration

18.175: Lecture 3 Integration 18.175: Lecture 3 Scott Sheffield MIT Outline Outline Recall definitions Probability space is triple (Ω, F, P) where Ω is sample space, F is set of events (the σ-algebra) and P : F [0, 1] is the probability

More information

Solution for Problem 7.1. We argue by contradiction. If the limit were not infinite, then since τ M (ω) is nondecreasing we would have

Solution for Problem 7.1. We argue by contradiction. If the limit were not infinite, then since τ M (ω) is nondecreasing we would have 362 Problem Hints and Solutions sup g n (ω, t) g(ω, t) sup g(ω, s) g(ω, t) µ n (ω). t T s,t: s t 1/n By the uniform continuity of t g(ω, t) on [, T], one has for each ω that µ n (ω) as n. Two applications

More information

Summarizing Measured Data

Summarizing Measured Data Summarizing Measured Data 12-1 Overview Basic Probability and Statistics Concepts: CDF, PDF, PMF, Mean, Variance, CoV, Normal Distribution Summarizing Data by a Single Number: Mean, Median, and Mode, Arithmetic,

More information

Contents Ordered Fields... 2 Ordered sets and fields... 2 Construction of the Reals 1: Dedekind Cuts... 2 Metric Spaces... 3

Contents Ordered Fields... 2 Ordered sets and fields... 2 Construction of the Reals 1: Dedekind Cuts... 2 Metric Spaces... 3 Analysis Math Notes Study Guide Real Analysis Contents Ordered Fields 2 Ordered sets and fields 2 Construction of the Reals 1: Dedekind Cuts 2 Metric Spaces 3 Metric Spaces 3 Definitions 4 Separability

More information

Math 710 Homework 1. Austin Mohr September 2, 2010

Math 710 Homework 1. Austin Mohr September 2, 2010 Math 710 Homework 1 Austin Mohr September 2, 2010 1 For the following random experiments, describe the sample space Ω For each experiment, describe also two subsets (events) that might be of interest,

More information

Statistics for Managers using Microsoft Excel 6 th Edition

Statistics for Managers using Microsoft Excel 6 th Edition Statistics for Managers using Microsoft Excel 6 th Edition Chapter 3 Numerical Descriptive Measures 3-1 Learning Objectives In this chapter, you learn: To describe the properties of central tendency, variation,

More information

Chapter 1 Statistical Reasoning Why statistics? Section 1.1 Basics of Probability Theory

Chapter 1 Statistical Reasoning Why statistics? Section 1.1 Basics of Probability Theory Chapter 1 Statistical Reasoning Why statistics? Uncertainty of nature (weather, earth movement, etc. ) Uncertainty in observation/sampling/measurement Variability of human operation/error imperfection

More information

Part III. 10 Topological Space Basics. Topological Spaces

Part III. 10 Topological Space Basics. Topological Spaces Part III 10 Topological Space Basics Topological Spaces Using the metric space results above as motivation we will axiomatize the notion of being an open set to more general settings. Definition 10.1.

More information

Probability and Measure

Probability and Measure Part II Year 2018 2017 2016 2015 2014 2013 2012 2011 2010 2009 2008 2007 2006 2005 2018 84 Paper 4, Section II 26J Let (X, A) be a measurable space. Let T : X X be a measurable map, and µ a probability

More information

Part V. 17 Introduction: What are measures and why measurable sets. Lebesgue Integration Theory

Part V. 17 Introduction: What are measures and why measurable sets. Lebesgue Integration Theory Part V 7 Introduction: What are measures and why measurable sets Lebesgue Integration Theory Definition 7. (Preliminary). A measure on a set is a function :2 [ ] such that. () = 2. If { } = is a finite

More information

DS-GA 1002 Lecture notes 2 Fall Random variables

DS-GA 1002 Lecture notes 2 Fall Random variables DS-GA 12 Lecture notes 2 Fall 216 1 Introduction Random variables Random variables are a fundamental tool in probabilistic modeling. They allow us to model numerical quantities that are uncertain: the

More information

The expected value E[X] of discrete random variable X is defined by. xp X (x), (6.1) E[X] =

The expected value E[X] of discrete random variable X is defined by. xp X (x), (6.1) E[X] = Chapter 6 Meeting Expectations When a large collection of data is gathered, one is typically interested not necessarily in every individual data point, but rather in certain descriptive quantities such

More information

First In-Class Exam Solutions Math 410, Professor David Levermore Monday, 1 October 2018

First In-Class Exam Solutions Math 410, Professor David Levermore Monday, 1 October 2018 First In-Class Exam Solutions Math 40, Professor David Levermore Monday, October 208. [0] Let {b k } k N be a sequence in R and let A be a subset of R. Write the negations of the following assertions.

More information

Sequences. Chapter 3. n + 1 3n + 2 sin n n. 3. lim (ln(n + 1) ln n) 1. lim. 2. lim. 4. lim (1 + n)1/n. Answers: 1. 1/3; 2. 0; 3. 0; 4. 1.

Sequences. Chapter 3. n + 1 3n + 2 sin n n. 3. lim (ln(n + 1) ln n) 1. lim. 2. lim. 4. lim (1 + n)1/n. Answers: 1. 1/3; 2. 0; 3. 0; 4. 1. Chapter 3 Sequences Both the main elements of calculus (differentiation and integration) require the notion of a limit. Sequences will play a central role when we work with limits. Definition 3.. A Sequence

More information

The main results about probability measures are the following two facts:

The main results about probability measures are the following two facts: Chapter 2 Probability measures The main results about probability measures are the following two facts: Theorem 2.1 (extension). If P is a (continuous) probability measure on a field F 0 then it has a

More information

Chapter 11 - Sequences and Series

Chapter 11 - Sequences and Series Calculus and Analytic Geometry II Chapter - Sequences and Series. Sequences Definition. A sequence is a list of numbers written in a definite order, We call a n the general term of the sequence. {a, a

More information

Set, functions and Euclidean space. Seungjin Han

Set, functions and Euclidean space. Seungjin Han Set, functions and Euclidean space Seungjin Han September, 2018 1 Some Basics LOGIC A is necessary for B : If B holds, then A holds. B A A B is the contraposition of B A. A is sufficient for B: If A holds,

More information

Statistical Data Analysis

Statistical Data Analysis DS-GA 0 Lecture notes 8 Fall 016 1 Descriptive statistics Statistical Data Analysis In this section we consider the problem of analyzing a set of data. We describe several techniques for visualizing the

More information

Statistics for Economists. Lectures 3 & 4

Statistics for Economists. Lectures 3 & 4 Statistics for Economists Lectures 3 & 4 Asrat Temesgen Stockholm University 1 CHAPTER 2- Discrete Distributions 2.1. Random variables of the Discrete Type Definition 2.1.1: Given a random experiment with

More information

Standard forms for writing numbers

Standard forms for writing numbers Standard forms for writing numbers In order to relate the abstract mathematical descriptions of familiar number systems to the everyday descriptions of numbers by decimal expansions and similar means,

More information

Construction of a general measure structure

Construction of a general measure structure Chapter 4 Construction of a general measure structure We turn to the development of general measure theory. The ingredients are a set describing the universe of points, a class of measurable subsets along

More information

1 + lim. n n+1. f(x) = x + 1, x 1. and we check that f is increasing, instead. Using the quotient rule, we easily find that. 1 (x + 1) 1 x (x + 1) 2 =

1 + lim. n n+1. f(x) = x + 1, x 1. and we check that f is increasing, instead. Using the quotient rule, we easily find that. 1 (x + 1) 1 x (x + 1) 2 = Chapter 5 Sequences and series 5. Sequences Definition 5. (Sequence). A sequence is a function which is defined on the set N of natural numbers. Since such a function is uniquely determined by its values

More information

Expectation of Random Variables

Expectation of Random Variables 1 / 19 Expectation of Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 13, 2015 2 / 19 Expectation of Discrete

More information

Theorems. Theorem 1.11: Greatest-Lower-Bound Property. Theorem 1.20: The Archimedean property of. Theorem 1.21: -th Root of Real Numbers

Theorems. Theorem 1.11: Greatest-Lower-Bound Property. Theorem 1.20: The Archimedean property of. Theorem 1.21: -th Root of Real Numbers Page 1 Theorems Wednesday, May 9, 2018 12:53 AM Theorem 1.11: Greatest-Lower-Bound Property Suppose is an ordered set with the least-upper-bound property Suppose, and is bounded below be the set of lower

More information

Notes 1 : Measure-theoretic foundations I

Notes 1 : Measure-theoretic foundations I Notes 1 : Measure-theoretic foundations I Math 733-734: Theory of Probability Lecturer: Sebastien Roch References: [Wil91, Section 1.0-1.8, 2.1-2.3, 3.1-3.11], [Fel68, Sections 7.2, 8.1, 9.6], [Dur10,

More information

MORE ON CONTINUOUS FUNCTIONS AND SETS

MORE ON CONTINUOUS FUNCTIONS AND SETS Chapter 6 MORE ON CONTINUOUS FUNCTIONS AND SETS This chapter can be considered enrichment material containing also several more advanced topics and may be skipped in its entirety. You can proceed directly

More information

We denote the space of distributions on Ω by D ( Ω) 2.

We denote the space of distributions on Ω by D ( Ω) 2. Sep. 1 0, 008 Distributions Distributions are generalized functions. Some familiarity with the theory of distributions helps understanding of various function spaces which play important roles in the study

More information

Math 324 Summer 2012 Elementary Number Theory Notes on Mathematical Induction

Math 324 Summer 2012 Elementary Number Theory Notes on Mathematical Induction Math 4 Summer 01 Elementary Number Theory Notes on Mathematical Induction Principle of Mathematical Induction Recall the following axiom for the set of integers. Well-Ordering Axiom for the Integers If

More information

Chapter 4.notebook. August 30, 2017

Chapter 4.notebook. August 30, 2017 Sep 1 7:53 AM Sep 1 8:21 AM Sep 1 8:21 AM 1 Sep 1 8:23 AM Sep 1 8:23 AM Sep 1 8:23 AM SOCS When describing a distribution, make sure to always tell about three things: shape, outliers, center, and spread

More information

3.1 Measure of Center

3.1 Measure of Center 3.1 Measure of Center Calculate the mean for a given data set Find the median, and describe why the median is sometimes preferable to the mean Find the mode of a data set Describe how skewness affects

More information

Spring 2014 Advanced Probability Overview. Lecture Notes Set 1: Course Overview, σ-fields, and Measures

Spring 2014 Advanced Probability Overview. Lecture Notes Set 1: Course Overview, σ-fields, and Measures 36-752 Spring 2014 Advanced Probability Overview Lecture Notes Set 1: Course Overview, σ-fields, and Measures Instructor: Jing Lei Associated reading: Sec 1.1-1.4 of Ash and Doléans-Dade; Sec 1.1 and A.1

More information

Random Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R

Random Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R In probabilistic models, a random variable is a variable whose possible values are numerical outcomes of a random phenomenon. As a function or a map, it maps from an element (or an outcome) of a sample

More information

Examples of Dual Spaces from Measure Theory

Examples of Dual Spaces from Measure Theory Chapter 9 Examples of Dual Spaces from Measure Theory We have seen that L (, A, µ) is a Banach space for any measure space (, A, µ). We will extend that concept in the following section to identify an

More information

Brief Review of Probability

Brief Review of Probability Maura Department of Economics and Finance Università Tor Vergata Outline 1 Distribution Functions Quantiles and Modes of a Distribution 2 Example 3 Example 4 Distributions Outline Distribution Functions

More information

4 Sums of Independent Random Variables

4 Sums of Independent Random Variables 4 Sums of Independent Random Variables Standing Assumptions: Assume throughout this section that (,F,P) is a fixed probability space and that X 1, X 2, X 3,... are independent real-valued random variables

More information

MATH 117 LECTURE NOTES

MATH 117 LECTURE NOTES MATH 117 LECTURE NOTES XIN ZHOU Abstract. This is the set of lecture notes for Math 117 during Fall quarter of 2017 at UC Santa Barbara. The lectures follow closely the textbook [1]. Contents 1. The set

More information

Product measure and Fubini s theorem

Product measure and Fubini s theorem Chapter 7 Product measure and Fubini s theorem This is based on [Billingsley, Section 18]. 1. Product spaces Suppose (Ω 1, F 1 ) and (Ω 2, F 2 ) are two probability spaces. In a product space Ω = Ω 1 Ω

More information

Stat 5101 Lecture Slides Deck 4. Charles J. Geyer School of Statistics University of Minnesota

Stat 5101 Lecture Slides Deck 4. Charles J. Geyer School of Statistics University of Minnesota Stat 5101 Lecture Slides Deck 4 Charles J. Geyer School of Statistics University of Minnesota 1 Existence of Integrals Just from the definition of integral as area under the curve, the integral b g(x)

More information

6.1 Moment Generating and Characteristic Functions

6.1 Moment Generating and Characteristic Functions Chapter 6 Limit Theorems The power statistics can mostly be seen when there is a large collection of data points and we are interested in understanding the macro state of the system, e.g., the average,

More information

Chapter 1. Sets and probability. 1.3 Probability space

Chapter 1. Sets and probability. 1.3 Probability space Random processes - Chapter 1. Sets and probability 1 Random processes Chapter 1. Sets and probability 1.3 Probability space 1.3 Probability space Random processes - Chapter 1. Sets and probability 2 Probability

More information

TABLE OF CONTENTS CHAPTER 1 COMBINATORIAL PROBABILITY 1

TABLE OF CONTENTS CHAPTER 1 COMBINATORIAL PROBABILITY 1 TABLE OF CONTENTS CHAPTER 1 COMBINATORIAL PROBABILITY 1 1.1 The Probability Model...1 1.2 Finite Discrete Models with Equally Likely Outcomes...5 1.2.1 Tree Diagrams...6 1.2.2 The Multiplication Principle...8

More information

Exercises for Unit VI (Infinite constructions in set theory)

Exercises for Unit VI (Infinite constructions in set theory) Exercises for Unit VI (Infinite constructions in set theory) VI.1 : Indexed families and set theoretic operations (Halmos, 4, 8 9; Lipschutz, 5.3 5.4) Lipschutz : 5.3 5.6, 5.29 5.32, 9.14 1. Generalize

More information

Numerical Measures of Central Tendency

Numerical Measures of Central Tendency ҧ Numerical Measures of Central Tendency The central tendency of the set of measurements that is, the tendency of the data to cluster, or center, about certain numerical values; usually the Mean, Median

More information

Math 180A. Lecture 16 Friday May 7 th. Expectation. Recall the three main probability density functions so far (1) Uniform (2) Exponential.

Math 180A. Lecture 16 Friday May 7 th. Expectation. Recall the three main probability density functions so far (1) Uniform (2) Exponential. Math 8A Lecture 6 Friday May 7 th Epectation Recall the three main probability density functions so far () Uniform () Eponential (3) Power Law e, ( ), Math 8A Lecture 6 Friday May 7 th Epectation Eample

More information

Statistics 3657 : Moment Generating Functions

Statistics 3657 : Moment Generating Functions Statistics 3657 : Moment Generating Functions A useful tool for studying sums of independent random variables is generating functions. course we consider moment generating functions. In this Definition

More information

A matrix over a field F is a rectangular array of elements from F. The symbol

A matrix over a field F is a rectangular array of elements from F. The symbol Chapter MATRICES Matrix arithmetic A matrix over a field F is a rectangular array of elements from F The symbol M m n (F ) denotes the collection of all m n matrices over F Matrices will usually be denoted

More information

Class 11 Maths Chapter 15. Statistics

Class 11 Maths Chapter 15. Statistics 1 P a g e Class 11 Maths Chapter 15. Statistics Statistics is the Science of collection, organization, presentation, analysis and interpretation of the numerical data. Useful Terms 1. Limit of the Class

More information

Metric Spaces and Topology

Metric Spaces and Topology Chapter 2 Metric Spaces and Topology From an engineering perspective, the most important way to construct a topology on a set is to define the topology in terms of a metric on the set. This approach underlies

More information

Simple Integer Recourse Models: Convexity and Convex Approximations

Simple Integer Recourse Models: Convexity and Convex Approximations Simple Integer Recourse Models: Convexity and Convex Approximations Willem K. Klein Haneveld Leen Stougie Maarten H. van der Vlerk November 8, 005 We consider the objective function of a simple recourse

More information

convergence theorem in abstract set up. Our proof produces a positive integrable function required unlike other known

convergence theorem in abstract set up. Our proof produces a positive integrable function required unlike other known https://sites.google.com/site/anilpedgaonkar/ profanilp@gmail.com 218 Chapter 5 Convergence and Integration In this chapter we obtain convergence theorems. Convergence theorems will apply to various types

More information

Lecture 11. Probability Theory: an Overveiw

Lecture 11. Probability Theory: an Overveiw Math 408 - Mathematical Statistics Lecture 11. Probability Theory: an Overveiw February 11, 2013 Konstantin Zuev (USC) Math 408, Lecture 11 February 11, 2013 1 / 24 The starting point in developing the

More information

STATISTICS SYLLABUS UNIT I

STATISTICS SYLLABUS UNIT I STATISTICS SYLLABUS UNIT I (Probability Theory) Definition Classical and axiomatic approaches.laws of total and compound probability, conditional probability, Bayes Theorem. Random variable and its distribution

More information

8 Laws of large numbers

8 Laws of large numbers 8 Laws of large numbers 8.1 Introduction We first start with the idea of standardizing a random variable. Let X be a random variable with mean µ and variance σ 2. Then Z = (X µ)/σ will be a random variable

More information

FIXED POINT ITERATIONS

FIXED POINT ITERATIONS FIXED POINT ITERATIONS MARKUS GRASMAIR 1. Fixed Point Iteration for Non-linear Equations Our goal is the solution of an equation (1) F (x) = 0, where F : R n R n is a continuous vector valued mapping in

More information

Expectation, variance and moments

Expectation, variance and moments Expectation, variance and moments John Appleby Contents Expectation and variance Examples 3 Moments and the moment generating function 4 4 Examples of moment generating functions 5 5 Concluding remarks

More information

STAT/MATH 395 PROBABILITY II

STAT/MATH 395 PROBABILITY II STAT/MATH 395 PROBABILITY II Chapter 6 : Moment Functions Néhémy Lim 1 1 Department of Statistics, University of Washington, USA Winter Quarter 2016 of Common Distributions Outline 1 2 3 of Common Distributions

More information

Product measures, Tonelli s and Fubini s theorems For use in MAT4410, autumn 2017 Nadia S. Larsen. 17 November 2017.

Product measures, Tonelli s and Fubini s theorems For use in MAT4410, autumn 2017 Nadia S. Larsen. 17 November 2017. Product measures, Tonelli s and Fubini s theorems For use in MAT4410, autumn 017 Nadia S. Larsen 17 November 017. 1. Construction of the product measure The purpose of these notes is to prove the main

More information

Limiting Distributions

Limiting Distributions Limiting Distributions We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the

More information

Lectures 2 3 : Wigner s semicircle law

Lectures 2 3 : Wigner s semicircle law Fall 009 MATH 833 Random Matrices B. Való Lectures 3 : Wigner s semicircle law Notes prepared by: M. Koyama As we set up last wee, let M n = [X ij ] n i,j=1 be a symmetric n n matrix with Random entries

More information

1.1. MEASURES AND INTEGRALS

1.1. MEASURES AND INTEGRALS CHAPTER 1: MEASURE THEORY In this chapter we define the notion of measure µ on a space, construct integrals on this space, and establish their basic properties under limits. The measure µ(e) will be defined

More information

M4L5. Expectation and Moments of Functions of Random Variable

M4L5. Expectation and Moments of Functions of Random Variable M4L5 Expectation and Moments of Functions of Random Variable 1. Introduction This lecture is a continuation of previous lecture, elaborating expectations, moments and moment generating functions of the

More information

Hartogs Theorem: separate analyticity implies joint Paul Garrett garrett/

Hartogs Theorem: separate analyticity implies joint Paul Garrett  garrett/ (February 9, 25) Hartogs Theorem: separate analyticity implies joint Paul Garrett garrett@math.umn.edu http://www.math.umn.edu/ garrett/ (The present proof of this old result roughly follows the proof

More information