Lecture 3: Expected Value. These integrals are taken over all of Ω. If we wish to integrate over a measurable subset A Ω, we will write

Similar documents
Lectures 22-23: Conditional Expectations

Graduate Probability Theory

18.175: Lecture 3 Integration

1 Probability space and random variables

0.1 Uniform integrability

4 Expectation & the Lebesgue Theorems

If Y and Y 0 satisfy (1-2), then Y = Y 0 a.s.

(1) Consider the space S consisting of all continuous real-valued functions on the closed interval [0, 1]. For f, g S, define

ABSTRACT EXPECTATION

Lecture 21: Expectation of CRVs, Fatou s Lemma and DCT Integration of Continuous Random Variables

CLASSICAL PROBABILITY MODES OF CONVERGENCE AND INEQUALITIES

L p Spaces and Convexity

Lecture 2: Random Variables and Expectation

Probability and Measure

Conditional expectation

Advanced Probability

Lebesgue Integration: A non-rigorous introduction. What is wrong with Riemann integration?

Definition 6.1. A metric space (X, d) is complete if every Cauchy sequence tends to a limit in X.

MATH MEASURE THEORY AND FOURIER ANALYSIS. Contents

MATH 418: Lectures on Conditional Expectation

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

Appendix B: Inequalities Involving Random Variables and Their Expectations

Math 735: Stochastic Analysis

Exercise 1. Let f be a nonnegative measurable function. Show that. where ϕ is taken over all simple functions with ϕ f. k 1.

2 (Bonus). Let A X consist of points (x, y) such that either x or y is a rational number. Is A measurable? What is its Lebesgue measure?

Lecture 4: Conditional expectation and independence

Homework 11. Solutions

Summary of Real Analysis by Royden

3 (Due ). Let A X consist of points (x, y) such that either x or y is a rational number. Is A measurable? What is its Lebesgue measure?

Lecture 22: Variance and Covariance

CHAPTER 3: LARGE SAMPLE THEORY

7 Convergence in R d and in Metric Spaces

Math 328 Course Notes

f(x)dx = lim f n (x)dx.

Math 5051 Measure Theory and Functional Analysis I Homework Assignment 3

STOR 635 Notes (S13)

h(x) lim H(x) = lim Since h is nondecreasing then h(x) 0 for all x, and if h is discontinuous at a point x then H(x) > 0. Denote

Measure and Integration: Solutions of CW2

Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension. n=1

1. Probability Measure and Integration Theory in a Nutshell

Lecture 5: Expectation

Lectures on Integration. William G. Faris

X n D X lim n F n (x) = F (x) for all x C F. lim n F n(u) = F (u) for all u C F. (2)

Math 635: An Introduction to Brownian Motion and Stochastic Calculus

3 Integration and Expectation

Continuous Random Variables

THEOREMS, ETC., FOR MATH 515

d(x n, x) d(x n, x nk ) + d(x nk, x) where we chose any fixed k > N

MATH 217A HOMEWORK. P (A i A j ). First, the basis case. We make a union disjoint as follows: P (A B) = P (A) + P (A c B)

Part II Probability and Measure

A D VA N C E D P R O B A B I L - I T Y

Probability: Handout

Brownian Motion and Conditional Probability

Useful Probability Theorems

Exercises Measure Theoretic Probability

Probability and Random Processes

Lecture 2: Convex Sets and Functions

Probability Review. Yutian Li. January 18, Stanford University. Yutian Li (Stanford University) Probability Review January 18, / 27

4. Conditional risk measures and their robust representation

Math212a1413 The Lebesgue integral.

1. Stochastic Processes and filtrations

Midterm Examination. STA 205: Probability and Measure Theory. Wednesday, 2009 Mar 18, 2:50-4:05 pm

Problem set 1, Real Analysis I, Spring, 2015.

Section Integration of Nonnegative Measurable Functions

EE514A Information Theory I Fall 2013

Problem set 4, Real Analysis I, Spring, 2015.

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 9 10/2/2013. Conditional expectations, filtration and martingales

Real Analysis Notes. Thomas Goller

The main results about probability measures are the following two facts:

Three hours THE UNIVERSITY OF MANCHESTER. 24th January

Chapter 6. Integration. 1. Integrals of Nonnegative Functions. a j µ(e j ) (ca j )µ(e j ) = c X. and ψ =

Lebesgue s Differentiation Theorem via Maximal Functions

Continuity. Chapter 4

Math 832 Fall University of Wisconsin at Madison. Instructor: David F. Anderson

Notes on Measure, Probability and Stochastic Processes. João Lopes Dias

Notes 15 : UI Martingales

Lecture 3 : Random variables and their distributions

Continuity. Matt Rosenzweig

THE LINDEBERG-FELLER CENTRAL LIMIT THEOREM VIA ZERO BIAS TRANSFORMATION

1 Measurable Functions

Lecture 4 Lebesgue spaces and inequalities

1/12/05: sec 3.1 and my article: How good is the Lebesgue measure?, Math. Intelligencer 11(2) (1989),

Hints/Solutions for Homework 3

L p Functions. Given a measure space (X, µ) and a real number p [1, ), recall that the L p -norm of a measurable function f : X R is defined by

SOME QUESTIONS FOR MATH 766, SPRING Question 1. Let C([0, 1]) be the set of all continuous functions on [0, 1] endowed with the norm

REAL VARIABLES: PROBLEM SET 1. = x limsup E k

THE STONE-WEIERSTRASS THEOREM AND ITS APPLICATIONS TO L 2 SPACES

Lecture 4: Completion of a Metric Space

The reference [Ho17] refers to the course lecture notes by Ilkka Holopainen.

Concentration Inequalities

Lecture 5 - Information theory

Joint Distribution of Two or More Random Variables

Review of measure theory

Lebesgue measure and integration

Measure and integration

Measure Theory on Topological Spaces. Course: Prof. Tony Dorlas 2010 Typset: Cathal Ormond

Fundamental Inequalities, Convergence and the Optional Stopping Theorem for Continuous-Time Martingales

Probability Theory. Richard F. Bass

Probability inequalities 11

Probability and Measure

Transcription:

Lecture 3: Expected Value 1.) Definitions. If X 0 is a random variable on (Ω, F, P), then we define its expected value to be EX = XdP. Notice that this quantity may be. For general X, we say that EX exists if the difference EX = EX + EX is well-defined, which it will be if either EX + < or EX <. These integrals are taken over all of Ω. If we wish to integrate over a measurable subset A Ω, we will write E[X; A] X dp X1 A dp. Notice that EX inherits all of the properties of the Lebesgue integral. In particular, Thm. (1.3.1): uppose that X, Y 0 or E X, E Y <. Then 1. E[X + Y ] = EX + EY. 2. E[aX + b] = ae[x] + b for any a, b R. 3. If P{X Y } = 1, then EX EY. A 2.) Inequalities: Jensen s inequality. uppose that φ : R R is convex, i.e., φ(λx + (1 λ)y) λφ(x) + (1 λ)φ(y) for all λ (0, 1) and all x, y R. Then φ (EX) E[φ(X)]. Examples: EX E X, (EX) 2 EX 2. Hölder s inequality: If p, q [1, ] with 1/p + 1/q = 1, then E XY X p Y q, where X r (E X r ) 1/r and X inf{m : P( X > M) = 0}. 1

X is called the essential supremum of X and satisfies the inequality X sup{ X(ω) : ω Ω}. Also, notice that the case p = q = 1/2 is the Cauchy-chwarz inequality: E XY (E[X 2 ] E[Y 2 ]). Chebyshev s inequality: uppose that φ : R R is non-negative, let A A and set i A = inf{φ(y) : y A}. Then i A P(X A) E[φ(X); A] E[φ(X)]. Proof: The result follows by taking expectations of the inequality i A 1 A (X) φ(x)1 A (X) φ(x). If φ(x) = x 2 and A = {x : x a}, then we obtain Markov s inequality: a 2 P{ X a} E[X 2 ]. 3.) Integration of limits: We are interested in conditions that guarantee that if X n X, then EX n EX. The following example shows that this does not hold in general. Example: Take Ω = (0, 1), F = the Borel sets and P = Lebesgue measure on (0, 1). X n = n1 (0,1/n), then X n X 0, but EX n = 1 > 0 = EX. If We begin by recalling three classical results from analysis. Fatou s Lemma: If X n 0, then [ ] E lim inf X n lim inf E[X n]. Monotone Convergence Theorem: If 0 X n X, then EX n EX. Proof: ince X n X for all n, we know that lim sup EX n EX. However, since X = lim inf X n, Fatou s Lemma implies that EX lim inf EX n. Combining these two results shows that EX = lim EX n. 2

Dominated Convergence Theorem: If X n X a.s. and X n Y for all n, where EY <, then EX n EX. The special case where Y is constant is called the bounded convergence theorem. The following theorem can handle some cases that are not covered by either the monotone or the dominated convergence theorems. Theorem (1.3.8): uppose that X n X a.s. Let g, h be continuous functions such that 1. g 0 and g(x) as x ; 2. h(x) /g(x) 0 as x ; 3. E[g(X n )] K < for all n. Then E[h(X n )] E[h(X)]. Proof: By subtracting a constant from h, we can assume wlog that h(0) = 0. Choose M so that P( X = M) = 0 and g(x) > 0 whenever x M. Given a random variable Y, let Ȳ = Y 1 ( Y M). Then X n X a.s. Indeed, on the set X < M, we have that X n < M for all n sufficiently large and so X n = X n X = X, while on the set X > M, we have that X n > M for all n sufficiently large and so X n = 0 X = 0. ince h( X n ) is bounded and h is continuous, the bounded convergence theorem implies that To control the truncation error, let E[h( X n )] E[h( X)]. ɛ M sup{ h(x) /g(x) : x M}. and observe that for any random variable Y we have ( ) E[h(Ȳ )] E[h(Y )] E h(ȳ ) h(y ) = E[ h(y ) ; Y > M] ɛ M E[g(Y )]. Taking Y = X n in ( ) and using condition (3) in the theorem, we have Eh( X n ) Eh(X n ) Kɛ M. To estimate the remaining truncation error, notice that because g 0 and g is continuous, Fatou s lemma implies that E[g(X)] lim inf E[g(X n)] K. Then, taking Y = X in ( ) gives Eh( X) Eh(X) Kɛ M. 3

Finally, by the triangle inequality, we have Eh(X n ) Eh(X) E[h(X n )] E[h( X n )] + E[h( X n )] E[h( X)] + E[h( X)] E[h(X)]. Letting n, we obtain lim sup E[h(Xn )] E[h(X)] 2KɛM which can be made arbitrarily close to 0 since ɛ M 0 as M. Corollary: uppose that X n X a.s. and that there exists a K < and a p > 1 such that E[X p n] K for all n 1. Then EX n EX. 4.) Computing Expected Values Change of Variables Formula: Let X be a random variable with distribution µ on (, ). If f is a measurable function from (, ) to (R, R) such that f 0 or E f(x) <, then E[f(X)] = f(y) µ(dy). Proof: We use the approach that Williams calls the standard machine. Case 1: Indicator Functions. If B and f = 1 B, then E[1 B (X)] = P(X B) = µ(b) = 1 B (y) µ(dy). Case 2: imple Functions. Let f(x) = n i=1 c i1 Bi (x), where c i R and B i. Then Case 1 combined with the linearity of both expectations and integration shows that E[f(X)] = = n c i E[1 Bi (X)] i=1 n c i i=1 1 Bi (y) µ(dy) = f(y) µ(dy). Case 3: Nonnegative Functions. If f 0 and we let f n (x) = ( 2 n f(x) /2 n ) n, then f n is a simple function and f n f as n. Combining Case 2 with the monotone convergence theorem gives E[f(X)] = lim E[f(X n)] = lim f n (y)µ(dy) = f(y)µ(dy). 4

Case 4: Integrable Functions. Write f(x) = f + (x) f (x) and note that the integrability of f implies that E[f + (X)] and E[f (X)] are finite. Using Case 3 gives Ef(X) = E[f + (X)] E[f (X)] = f + (y) µ(dy) f (y) µ(dy) = f(y) µ(dy). 5