Advanced Probability

Similar documents
A D VA N C E D P R O B A B I L - I T Y

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 9 10/2/2013. Conditional expectations, filtration and martingales

4 Expectation & the Lebesgue Theorems

MATH 418: Lectures on Conditional Expectation

If Y and Y 0 satisfy (1-2), then Y = Y 0 a.s.

Brownian Motion and Conditional Probability

18.175: Lecture 3 Integration

Lecture 2: Random Variables and Expectation

Lectures 22-23: Conditional Expectations

Conditional expectation

Lecture 22: Variance and Covariance

JUSTIN HARTMANN. F n Σ.

1. Probability Measure and Integration Theory in a Nutshell

MATH MEASURE THEORY AND FOURIER ANALYSIS. Contents

THEOREMS, ETC., FOR MATH 515

Notes on Measure, Probability and Stochastic Processes. João Lopes Dias

Notes 1 : Measure-theoretic foundations I

CONVERGENCE OF RANDOM SERIES AND MARTINGALES

Lebesgue Integration: A non-rigorous introduction. What is wrong with Riemann integration?

Lecture Notes for MA 623 Stochastic Processes. Ionut Florescu. Stevens Institute of Technology address:

Lecture 3: Expected Value. These integrals are taken over all of Ω. If we wish to integrate over a measurable subset A Ω, we will write

Probability: Handout

Notes 13 : Conditioning

ABSTRACT EXPECTATION

Lecture 6 Basic Probability

We will briefly look at the definition of a probability space, probability measures, conditional probability and independence of probability events.

Errata for FIRST edition of A First Look at Rigorous Probability

7 Convergence in R d and in Metric Spaces

1. Stochastic Processes and filtrations

Fundamental Inequalities, Convergence and the Optional Stopping Theorem for Continuous-Time Martingales

Probability and Measure

Probability and Measure

MATH/STAT 235A Probability Theory Lecture Notes, Fall 2013

PROBABILITY THEORY II

Martingale Theory and Applications

Lecture 21: Expectation of CRVs, Fatou s Lemma and DCT Integration of Continuous Random Variables

Reminder Notes for the Course on Measures on Topological Spaces

4 Expectation & the Lebesgue Theorems

Preliminaries to Stochastic Analysis Autumn 2011 version. Xue-Mei Li The University of Warwick

(U) =, if 0 U, 1 U, (U) = X, if 0 U, and 1 U. (U) = E, if 0 U, but 1 U. (U) = X \ E if 0 U, but 1 U. n=1 A n, then A M.

MATH & MATH FUNCTIONS OF A REAL VARIABLE EXERCISES FALL 2015 & SPRING Scientia Imperii Decus et Tutamen 1

Elementary Probability. Exam Number 38119

Exercises Measure Theoretic Probability

5 Measure theory II. (or. lim. Prove the proposition. 5. For fixed F A and φ M define the restriction of φ on F by writing.

3 Integration and Expectation

Spring 2014 Advanced Probability Overview. Lecture Notes Set 1: Course Overview, σ-fields, and Measures

Real Analysis Notes. Thomas Goller

Part II Probability and Measure

Probability Theory. Richard F. Bass

0.1 Uniform integrability

1 Measurable Functions

Part V. 17 Introduction: What are measures and why measurable sets. Lebesgue Integration Theory

2 (Bonus). Let A X consist of points (x, y) such that either x or y is a rational number. Is A measurable? What is its Lebesgue measure?

Measures. 1 Introduction. These preliminary lecture notes are partly based on textbooks by Athreya and Lahiri, Capinski and Kopp, and Folland.

Notes 15 : UI Martingales

Independent random variables

Summary of Real Analysis by Royden

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 3 9/10/2008 CONDITIONING AND INDEPENDENCE

Stat 451: Solutions to Assignment #1

STOCHASTIC MODELS FOR WEB 2.0

Dynkin (λ-) and π-systems; monotone classes of sets, and of functions with some examples of application (mainly of a probabilistic flavor)

Lecture 3 : Random variables and their distributions

Random Process Lecture 1. Fundamentals of Probability

REAL AND COMPLEX ANALYSIS

1 Stochastic Dynamic Programming

Lecture 12: Multiple Random Variables and Independence

HILBERT SPACES AND THE RADON-NIKODYM THEOREM. where the bar in the first equation denotes complex conjugation. In either case, for any x V define

Section 2: Classes of Sets

Measure Theoretic Probability. P.J.C. Spreij

3 (Due ). Let A X consist of points (x, y) such that either x or y is a rational number. Is A measurable? What is its Lebesgue measure?

Integration on Measure Spaces

Measure and integration

Lecture 2. We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales.

2 Probability, random elements, random sets

(1) Consider the space S consisting of all continuous real-valued functions on the closed interval [0, 1]. For f, g S, define

ON THE EQUIVALENCE OF CONGLOMERABILITY AND DISINTEGRABILITY FOR UNBOUNDED RANDOM VARIABLES

Product measures, Tonelli s and Fubini s theorems For use in MAT4410, autumn 2017 Nadia S. Larsen. 17 November 2017.

MATHS 730 FC Lecture Notes March 5, Introduction

An Introduction to Stochastic Calculus

Part III Advanced Probability

18.175: Lecture 2 Extension theorems, random variables, distributions

2 Lebesgue integration

Measure Theory on Topological Spaces. Course: Prof. Tony Dorlas 2010 Typset: Cathal Ormond

Projection-valued measures and spectral integrals

Math 735: Stochastic Analysis

2 Measure Theory. 2.1 Measures

Math 6810 (Probability) Fall Lecture notes

ABSTRACT INTEGRATION CHAPTER ONE

STOR 635 Notes (S13)

L p Spaces and Convexity

1 Probability space and random variables

Three hours THE UNIVERSITY OF MANCHESTER. 24th January

ADVANCED PROBABILITY: SOLUTIONS TO SHEET 1

1.4 Outer measures 10 CHAPTER 1. MEASURE

Lecture 11: Random Variables

Functional Analysis I

Spectral Theory, with an Introduction to Operator Means. William L. Green

Some basic elements of Probability Theory

Advanced Probability

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product

Transcription:

Advanced Probability Perla Sousi October 10, 2011 Contents 1 Conditional expectation 1 1.1 Discrete case.................................. 3 1.2 Existence and uniqueness............................ 3 1 Conditional expectation Let (Ω, F, P) be a probability space, i.e. Ω is a set, F is a σ-algebra on Ω and P is a probability measure on (Ω, F). Definition 1.1. F is a σ-algebra on Ω if it satisfies: 1. Ω F 2. If A F, then also the complement is in F, i.e.,a c F. 3. If (A n ) n 1 is a collection of sets in F, then n=1a n F. Definition 1.2. P is a probability measure on (Ω, F) if it satisfies: 1. P : F [0, 1], i.e. it is a set function 2. P(Ω) = 1 and P( ) = 0 3. If (A n ) n 1 is a collection of pairwise disjoint sets in F, then P( n=1 A n) = n=1 P(A n). Let A, B F be two events with P(B) > 0. Then the conditional probability of A given the event B is defined by P(A B) P(A B) =. P(B) University of Cambridge, Cambridge, UK; p.sousi@statslab.cam.ac.uk 1

Definition 1.3. The Borel σ-algebra, B(R), is the σ-algebra generated by the open sets in R, i.e., it is the intersection of all σ-algebras containing the open sets of R. More formally, let O be the open sets of R, then B(R) = {E : E is a σ-algebra containing O}. Informally speaking, consider the open sets of R, do all possible operations, i.e., unions, intersections, complements. and take the smallest σ-algebra that you get. Definition 1.4. X is a random variable, i.e., a measurable function with respect to F, if X : Ω R is a function with the property that for all open sets V the inverse image X 1 (V ) F. Remark 1.5. If X is a random variable, then the collection of sets {B R : X 1 (B) F} is a σ-algebra (check!) and hence it must contain B(R). Let A F. The indicator function 1(A) is defined via { 1, if x A; 1(A)(x) = 1(x A) = 0, otherwise. Recall the definition of expectation. First for positive simple random variables, i.e., linear combinations of indicator random variables, we define [ n ] n E c i 1(A i ) := c i P(A i ), i=1 where c i are positive constants and A i are measurable events. Next, let X be a positive random variable. Then X is the increasing limit of positive simple variables. For example So we define i=1 X n (ω) = 2 n 2 n X(ω) n X(ω) as n. E[X] := lim E[X n ]. Finally, for a general random variable X, we can write X = X + X, where X + = max(x, 0) and X = max( X, 0) and we define E[X] := E[X + ] E[X ], if at least one of E[X + ], E[X ] is finite. E[ X ] <. A random variable X is called integrable, if 2

Let X be a random variable with E[ X ] <. Let A be an event in F with P(A) > 0. Then the conditional expectation of X given A is defined by E[X A] = E[X1(A)], P(A) Our goal is to extend the definition of conditional expectation to σ-algebras. So far we only have defined it for events and it was a number. Now, the conditional expectation is going to be a random variable, measurable with respect to the σ-algebra wrt which we are conditioning. 1.1 Discrete case Let X L 1. Let s start with a σ-algebra which is generated by a countable family of disjoint events (B i ) i I with i B i = Ω, i.e., G = σ(b i, i I). It is easy to check that G = { i J B i : J N}. The natural thing to do is to define a new random variable X = E[X G] as follows X = i I E[X B i ]1(B i ). What does this mean? Let ω Ω. Then X (ω) = i I E[X B i]1(ω B i ). Note that we use the convention that E[X B i ] = 0, if P(B i ) = 0 It is very easy to check that and integrable, since X is G measurable (1.1) E[ X ] i I E X1(B i ) = E[ X ] <, since X L 1. Let G G. Then it is straightforward to check that 1.2 Existence and uniqueness E[X1(G)] = E[X 1(G)]. (1.2) Before stating the existence and uniqueness theorem on conditional expectation, let us quickly recall the notion of an event happening almost surely (a.s.), the Monotone convergence theorem and L p spaces. Let A F. We will say that A happens a.s., if P(A) = 1. 3

Theorem 1.6. [Monotone convergence theorem] Let (X n ) n be random variables such that X n 0 for all n and X n X as n a.s. Then E[X n ] E[X] as n. Let p [1, ) and f a measurable function in (Ω, F, P). We define the norm f p = (E[ f p ]) 1/p and we denote by L p = L p (Ω, F, P) the set of measurable functions f with f p <. For p =, we let f = inf{λ : f λ a.e.} and L the set of measurable functions with f <. Formally, L p is the collection of equivalence classes, where two functions are equivalent if they are equal almost everywhere (a.e.). In practice, we will represent an element of L p by a function, but remember that equality in L p means equality a.e.. Theorem 1.7. The space (L 2, 2 ) is a Hilbert space with f, g = E[fg]. If H is a closed subspace, then for all f L 2, there exists a unique (in the sense of a.e.) g H such that f g 2 = inf h H f h 2 and g, f g = 0. Remark 1.8. We call g the orthogonal projection of f on H. Theorem 1.9. Let X be an integrable random variable and let G F be a σ-algebra. Then there exists a random variable Y such that: (a) Y is G-measurable; (b) Y is integrable and E[X1(A)] = E[Y 1(A)] for all A G. Moreover, if Y also satisfies (a) and (b), then Y = Y a.s.. We call Y (a version of ) the conditional expectation of X given G and write Y = E[X G] a.s.. In the case G = σ(g) for some random variable G, we also write Y = E[X G] a.s.. Remark 1.10. We could replace (b) in the statement of the theorem by requiring that for all bounded G-measurable random variables Z we have E[XZ] = E[Y Z]. Remark 1.11. In a later section we will show how to construct explicit versions of the conditional expectation in certain simple cases. In general, we have to live with the indirect approach provided by the theorem. 4

Proof of Theorem 1.9. (Uniqueness.) Suppose that both Y and Y satisfy (a) and (b). Then, clearly the event A = {Y > Y } G and by (b) we have E[(Y Y )1(A)] = E[X1(A)] E[X1(A)] = 0, hence we get that Y Y a.s. Similarly we can get Y Y a.s. (Existence.) We will prove existence in three steps. 1st step: Suppose that X L 2. The space L 2 (Ω, F, P) with inner product defined by U, V = E[UV ] is a Hilbert space by Theorem 1.7 and L 2 (Ω, G, P) is a closed subspace. (Remember that L 2 convergence implies convergence in probability and convergence in probability implies convergence a.s. along a subsequence (see for instance [1, A13.2])). Thus L 2 (F) = L 2 (G)+L 2 (G), and hence, we can write X as X = Y +Z, where Y L 2 (G) and Z L 2 (G). If we now set Y = E[X G], then (a) is clearly satisfied. Let A G. Then E[X1(A)] = E[Y1(A)] + E[Z1(A)] = E[Y1(A)], since E[Z1(A)] = 0. Note that from the above definition of conditional expectation for random variables in L 2, we get that since note that {Y < 0} G and if X 0, then Y = E[X G] 0 a.s., (1.3) E[X1(Y < 0)] = E[Y1(Y < 0)]. Notice that the left hand side is nonnegative, while the right hand side is non-positive, implying that P(Y < 0) = 0. 2nd step: Suppose that X 0. For each n we define the random variables X n = X n n, and hence X n L 2. Thus from the first part of the existence proof we have that for each n there exists a G-measurable random variable Y n satisfying for all A G E[Y n 1(A)] = E[(X n)1(a)]. (1.4) Since the sequence (X n ) n is increasing, from (1.3) we get that also (Y n ) n is increasing. If we now set Y = lim Y n, then clearly Y is G-measurable and by the monotone convergence theorem in (1.4) we get for all A G E[Y1(A)] = E[X1(A)], (1.5) since X n X, as n. In particular, if E[X] is finite, then E[Y ] is also finite. 3rd step: Finally, for a general random variable X L 1 (not necessarily positive) we can apply the above construction to X + = max(x, 0) and X = max( X, 0) and then E[X G] = E[X + G] E[X G] satisfies (a) and (b). 5

Remark 1.12. Note that the 2nd step of the above proof gives that if X 0, then there exists a G-measurable random variable Y such that for all A G, E[X1(A)] = E[Y1(A)], i.e., all the conditions of Theorem 1.9 are satisfied except for the integrability one. Definition 1.13. Sub-σ-algebras G 1, G 2,... of F are called independent, if whenever G i G i (i N) and i 1,...,i n are distinct, then P(G i1... G in ) = n P(G ik ). When we say that a random variable X is independent of a σ-algebra G, it means that σ(x) is independent of G. The following properties are immediate consequences of Theorem 1.9 and its proof. Proposition 1.14. Let X, Y L 1 (Ω, F, P) and let G F be a σ-algebra. Then 1. E [E[X G]] = E[X] 2. If X is G-measurable, then E[X G] = X a.s.. k=1 3. If X is independent of G, then E[X G] = E[X] a.s.. 4. If X 0, then E[X G] 0 a.s.. 5. For any α, β R we have E[αX + βy G] = αe[x G] + βe[y G] a.s.. 6. E[X G] E[ X G] a.s.. The basic convergence theorems for expectation have counterparts for conditional expectation. We first recall the theorems for expectation. Theorem 1.15. [Fatou s lemma] If X n 0 for all n, then E[lim inf n X n ] lim inf n E[X n ]. Theorem 1.16. [Dominated convergence theorem] If X n X and X n Y for all n a.s., for some integrable random variable Y, then E[X n ] E[X]. Theorem 1.17. [Jensen s inequality] Let X be an integrable random variable and let ϕ : R R be a convex function. Then E[ϕ(X)] ϕ(e[x]). 6

Proposition 1.18. Let G F be a σ-algebra and let X L 1 (Ω, F, P). 1. Conditional monotone convergence theorem: If (X n ) n 0 is an increasing sequence of nonnegative random variables with a.s. limit X, then E[X n G] ր E[X G] as n, a.s.. 2. Conditional Fatou s lemma: If X n 0 for all n, then [ ] E lim inf X n G lim inf E[X n G] a.s.. 3. Conditional dominated convergence theorem: If X n X and X n Y for all n a.s., for some integrable random variable Y, then lim E[X n G] = E[X G] a.s.. 4. Conditional Jensen s inequality: If X is integrable random and ϕ : R (, ] is a convex function such that either ϕ(x) is integrable or ϕ is non-negative, then E[ϕ(X) G] ϕ(e[x G]) a.s.. In particular, for all 1 p < E[X G] p X p. Proof. 1. Since X n ր X as n, we have that E[X n G] is an increasing sequence. Let Y = lim E[X n G]. We want to show that Y = E[X G] a.s.. Clearly Y is G-measurable, as an a.s. limit of G-measurable random variables. Also, by the monotone convergence theorem we have E[X1(A)] = lim E[X n 1(A)] = lim E[E[X n G]1(A)] = E[Y1(A)]. 2. The sequence inf k n X k is increasing in n and lim inf k n X k = lim inf X n. Thus, by the conditional monotone convergence theorem we get lim E[inf X k G] = E[lim inf X n G]. k n Clearly, E[inf k n X k G] inf k n E[X k G]. Passing to the limit gives the desired inequality. 3. Since X n +Y and Y X n are positive random variables for all n, applying conditional Fatou s lemma we get E[X + Y G] = E[lim inf(x n + Y ) G] lim inf E[Y X G] = E[lim inf(y X n ) G] lim inf Hence, we obtain that lim inf E[X n + Y G] and E[Y X n G]. E[X n G] E[X G] and lim sup E[X n G] E[X G]. 7

4. A convex function is the supremum of countably many affine functions: (see for instance [1, 6.6]) ϕ(x) = sup(a i x + b i ), x R. i So for all i we have E[ϕ(X) G] a i E[X G] + b i a.s. Now using the fact that the supremum is over a countable set we get that In particular, for 1 p <, E[ϕ(X) G] sup(a i E[X G] + b i ) = ϕ(e[x G]) a.s. i E[X G] p p = E[ E[X G] p ] E[E[ X p G]] = E[ X p ] = X p p. References [1] David Williams. Probability with martingales. Cambridge Mathematical Textbooks. Cambridge University Press, Cambridge, 1991. 8