Part II Probability and Measure

Similar documents
Probability and Measure

MATH MEASURE THEORY AND FOURIER ANALYSIS. Contents

MATHS 730 FC Lecture Notes March 5, Introduction

3 (Due ). Let A X consist of points (x, y) such that either x or y is a rational number. Is A measurable? What is its Lebesgue measure?

THEOREMS, ETC., FOR MATH 515

2 (Bonus). Let A X consist of points (x, y) such that either x or y is a rational number. Is A measurable? What is its Lebesgue measure?

Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension. n=1

Reminder Notes for the Course on Measures on Topological Spaces

Real Analysis Notes. Thomas Goller

Random Process Lecture 1. Fundamentals of Probability

Product measure and Fubini s theorem

Measure Theory & Integration

MTH 404: Measure and Integration

(1) Consider the space S consisting of all continuous real-valued functions on the closed interval [0, 1]. For f, g S, define

MATH & MATH FUNCTIONS OF A REAL VARIABLE EXERCISES FALL 2015 & SPRING Scientia Imperii Decus et Tutamen 1

18.175: Lecture 3 Integration

Lectures on Integration. William G. Faris

Analysis Comprehensive Exam Questions Fall 2008

The main results about probability measures are the following two facts:

Part III Advanced Probability

MATH 650. THE RADON-NIKODYM THEOREM

THEOREMS, ETC., FOR MATH 516

3. (a) What is a simple function? What is an integrable function? How is f dµ defined? Define it first

Probability Theory. Richard F. Bass

L p Spaces and Convexity

I. ANALYSIS; PROBABILITY

3 Integration and Expectation

STAT 7032 Probability Spring Wlodek Bryc

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

INTRODUCTION TO MEASURE THEORY AND LEBESGUE INTEGRATION

Real Analysis Problems

MATH 614 Dynamical Systems and Chaos Lecture 38: Ergodicity (continued). Mixing.

Infinite-dimensional Vector Spaces and Sequences

be the set of complex valued 2π-periodic functions f on R such that

Some basic elements of Probability Theory

Probability and Random Processes

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product

+ 2x sin x. f(b i ) f(a i ) < ɛ. i=1. i=1

Notation. General. Notation Description See. Sets, Functions, and Spaces. a b & a b The minimum and the maximum of a and b

A D VA N C E D P R O B A B I L - I T Y

Summary of Real Analysis by Royden

Probability and Measure: Alternative notes

Exercises Measure Theoretic Probability

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

Lebesgue Integration: A non-rigorous introduction. What is wrong with Riemann integration?

Review of measure theory

FACULTY OF SCIENCE FINAL EXAMINATION MATHEMATICS MATH 355. Analysis 4. Examiner: Professor S. W. Drury Date: Tuesday, April 18, 2006 INSTRUCTIONS

The Canonical Gaussian Measure on R

Weak convergence. Amsterdam, 13 November Leiden University. Limit theorems. Shota Gugushvili. Generalities. Criteria

Three hours THE UNIVERSITY OF MANCHESTER. 24th January

1 Independent increments

HILBERT SPACES AND THE RADON-NIKODYM THEOREM. where the bar in the first equation denotes complex conjugation. In either case, for any x V define

Solutions to Tutorial 11 (Week 12)

Probability and Measure

5 Measure theory II. (or. lim. Prove the proposition. 5. For fixed F A and φ M define the restriction of φ on F by writing.

Probability and Measure. November 27, 2017

ABSTRACT EXPECTATION

Lebesgue Integration on R n

If Y and Y 0 satisfy (1-2), then Y = Y 0 a.s.

STAT 7032 Probability. Wlodek Bryc

PROBLEMS. (b) (Polarization Identity) Show that in any inner product space

Advanced Analysis Qualifying Examination Department of Mathematics and Statistics University of Massachusetts. Tuesday, January 16th, 2018

1/12/05: sec 3.1 and my article: How good is the Lebesgue measure?, Math. Intelligencer 11(2) (1989),

Brownian Motion and Conditional Probability

Measure Theory on Topological Spaces. Course: Prof. Tony Dorlas 2010 Typset: Cathal Ormond

FACULTY OF SCIENCE FINAL EXAMINATION MATHEMATICS MATH 355. Analysis 4. Examiner: Professor S. W. Drury Date: Wednesday, April 18, 2007 INSTRUCTIONS

Tools from Lebesgue integration

6.2 Fubini s Theorem. (µ ν)(c) = f C (x) dµ(x). (6.2) Proof. Note that (X Y, A B, µ ν) must be σ-finite as well, so that.

Exercises Measure Theoretic Probability

1.1. MEASURES AND INTEGRALS

Lecture 4 Lebesgue spaces and inequalities

MEASURE AND INTEGRATION. Dietmar A. Salamon ETH Zürich

It follows from the above inequalities that for c C 1

It follows from the above inequalities that for c C 1

Let R be the line parameterized by x. Let f be a complex function on R that is integrable. The Fourier transform ˆf = F f is. e ikx f(x) dx. (1.

Bounded and continuous functions on a locally compact Hausdorff space and dual spaces

Lebesgue Measure on R n

Notes on Measure, Probability and Stochastic Processes. João Lopes Dias

Preliminaries to Stochastic Analysis Autumn 2011 version. Xue-Mei Li The University of Warwick

Measurable functions are approximately nice, even if look terrible.

ABSTRACT CONDITIONAL EXPECTATION IN L 2

Math 5051 Measure Theory and Functional Analysis I Homework Assignment 2

Class Notes for Math 921/922: Real Analysis, Instructor Mikil Foss

Product measures, Tonelli s and Fubini s theorems For use in MAT4410, autumn 2017 Nadia S. Larsen. 17 November 2017.

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Lecture 6 Basic Probability

Functional Analysis, Stein-Shakarchi Chapter 1

Real Analysis Qualifying Exam May 14th 2016

Measure Theory. John K. Hunter. Department of Mathematics, University of California at Davis

212a1214Daniell s integration theory.

LECTURE NOTES FOR , FALL Contents. Introduction. 1. Continuous functions

1 Measurable Functions

A PRIMER OF LEBESGUE INTEGRATION WITH A VIEW TO THE LEBESGUE-RADON-NIKODYM THEOREM

Math 832 Fall University of Wisconsin at Madison. Instructor: David F. Anderson

1 Probability space and random variables

Lebesgue Measure on R n

Probability and Measure

Advanced Analysis Qualifying Examination Department of Mathematics and Statistics University of Massachusetts. Monday, August 26, 2013

Functional Analysis. Franck Sueur Metric spaces Definitions Completeness Compactness Separability...

Measures. Chapter Some prerequisites. 1.2 Introduction

Transcription:

Part II Probability and Measure Theorems Based on lectures by J. Miller Notes taken by Dexter Chua Michaelmas 2016 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures. They are nowhere near accurate representations of what was actually lectured, and in particular, all errors are almost surely mine. Analysis II is essential Measure spaces, σ-algebras, π-systems and uniqueness of extension, statement *and proof* of Carathéodory s extension theorem. Construction of Lebesgue measure on R. The Borel σ-algebra of R. Existence of non-measurable subsets of R. Lebesgue-Stieltjes measures and probability distribution functions. Independence of events, independence of σ-algebras. The Borel Cantelli lemmas. Kolmogorov s zero-one law. [6] Measurable functions, random variables, independence of random variables. Construction of the integral, expectation. Convergence in measure and convergence almost everywhere. Fatou s lemma, monotone and dominated convergence, differentiation under the integral sign. Discussion of product measure and statement of Fubini s theorem. [6] Chebyshev s inequality, tail estimates. Jensen s inequality. Completeness of L p for 1 p. The Hölder and Minkowski inequalities, uniform integrability. [4] L 2 as a Hilbert space. Orthogonal projection, relation with elementary conditional probability. Variance and covariance. Gaussian random variables, the multivariate normal distribution. [2] The strong law of large numbers, proof for independent random variables with bounded fourth moments. Measure preserving transformations, Bernoulli shifts. Statements *and proofs* of maximal ergodic theorem and Birkhoff s almost everywhere ergodic theorem, proof of the strong law. [4] The Fourier transform of a finite measure, characteristic functions, uniqueness and inversion. Weak convergence, statement of Lévy s convergence theorem for characteristic functions. The central limit theorem. [2] 1

Contents II Probability and Measure (Theorems) Contents 0 Introduction 3 1 Measures 4 1.1 Measures............................... 4 1.2 Probability measures......................... 4 2 Measurable functions and random variables 5 2.1 Measurable functions......................... 5 2.2 Constructing new measures..................... 5 2.3 Random variables........................... 6 2.4 Convergence of measurable functions................ 6 2.5 Tail events............................... 7 3 Integration 8 3.1 Definition and basic properties................... 8 3.2 Integrals and limits.......................... 8 3.3 New measures from old....................... 9 3.4 Integration and differentiation.................... 9 3.5 Product measures and Fubini s theorem.............. 10 4 Inequalities and L p spaces 12 4.1 Four inequalities........................... 12 4.2 L p spaces............................... 12 4.3 Orthogonal projection in L 2..................... 13 4.4 Convergence in L 1 (P) and uniform integrability.......... 13 5 Fourier transform 14 5.1 The Fourier transform........................ 14 5.2 Convolutions............................. 14 5.3 Fourier inversion formula...................... 14 5.4 Fourier transform in L 2....................... 15 5.5 Properties of characteristic functions................ 15 5.6 Gaussian random variables..................... 15 6 Ergodic theory 17 6.1 Ergodic theorems........................... 17 7 Big theorems 18 7.1 The strong law of large numbers.................. 18 7.2 Central limit theorem........................ 18 2

0 Introduction II Probability and Measure (Theorems) 0 Introduction 3

1 Measures II Probability and Measure (Theorems) 1 Measures 1.1 Measures Proposition. A collection A is a σ-algebra if and only if it is both a π-system and a d-system. Lemma (Dynkin s π-system lemma). Let A be a π-system. Then any d-system which contains A contains σ(a). Theorem (Caratheodory extension theorem). Let A be a ring on E, and µ a countably additive set function on A. Then µ extends to a measure on the σ-algebra generated by A. Theorem. Suppose that µ 1, µ 2 are measures on (E, E) with µ 1 (E) = µ 2 (E) <. If A is a π-system with σ(a) = E, and µ 1 agrees with µ 2 on A, then µ 1 = µ 2. Theorem. There exists a unique Borel measure µ on R with µ([a, b]) = b a. Proposition. The Lebesgue measure is translation invariant, i.e. for all A B and x R, where µ(a + x) = µ(a) A + x = {y + x, y A}. Proposition. Let µ be a Borel measure on R that is translation invariant and µ([0, 1]) = 1. Then µ is the Lebesgue measure. 1.2 Probability measures Proposition. Events (A n ) are independent iff the σ-algebras σ(a n ) are independent. Theorem. Suppose A 1 and A 2 are π-systems in F. If P[A 1 A 2 ] = P[A 1 ]P[A 2 ] for all A 1 A 1 and A 2 A 2, then σ(a 1 ) and σ(a 2 ) are independent. Lemma (Borel Cantelli lemma). If P[A n ] <, then n P[A n i.o.] = 0. Lemma (Borel Cantelli lemma II). Let (A n ) be independent events. If P[A n ] =, then n P[A n i.o.] = 1. 4

2 Measurable functions and random variables II Probability and Measure (Theorems) 2 Measurable functions and random variables 2.1 Measurable functions Lemma. Let (E, E) and (G, G) be measurable spaces, and G = σ(q) for some Q. If f 1 (A) E for all A Q, then f is measurable. Proposition. Let f i : E F i be functions. Then {f i } are all measurable iff (f i ) : E F i is measurable, where the function (f i ) is defined by setting the ith component of (f i )(x) to be f i (x). Proposition. Let (E, E) be a measurable space. Let (f n : n N) be a sequence of non-negative measurable functions on E. Then the following are measurable: f 1 + f 2, f 1 f 2, max{f 1, f 2 }, min{f 1, f 2 }, inf f n, sup f n, n n lim inf f n, n lim sup f n. n The same is true with real replaced with non-negative, provided the new functions are real (i.e. not infinity). Theorem (Monotone class theorem). Let (E, E) be a measurable space, and A E be a π-system with σ(a) = E. Let V be a vector space of functions such that (i) The constant function 1 = 1 E is in V. (ii) The indicator functions 1 A V for all A A (iii) V is closed under bounded, monotone limits. More explicitly, if (f n ) is a bounded non-negative sequence in V, f n f (pointwise) and f is also bounded, then f V. Then V contains all bounded measurable functions. 2.2 Constructing new measures Lemma. Let g : R R be non-constant, non-decreasing and right continuous. We set g(± ) = lim g(x). x ± We set I = (g( ), g( )). Since g is non-constant, this is non-empty. Then there is a non-decreasing, left continuous function f : I R such that for all x I and y R, we have x g(y) f(x) y. Thus, taking the negation of this, we have Explicitly, for x I, we define x > g(y) f(x) > y. f(x) = inf{y R : x g(y)}. Theorem. Let g : R R be non-constant, non-decreasing and right continuous. Then there exists a unique Radon measure dg on B such that dg((a, b]) = g(b) g(a). Moreover, we obtain all non-zero Radon measures on R in this way. 5

2 Measurable functions and random variables II Probability and Measure (Theorems) 2.3 Random variables Proposition. We have F X (x) { 0 x 1 x +. Also, F X (x) is non-decreasing and right-continuous. Proposition. Let F be any distribution function. Then there exists a probability space (Ω, F, P) and a random variable X such that F X = F. Proposition. Two real-valued random variables X, Y are independent iff P[X x, Y y] = P[X x]p[y y]. More generally, if (X n ) is a sequence of real-valued random variables, then they are independent iff for all n and x j. Proposition. Let P[x 1 x 1,, x n x n ] = n P[X j x j ] j=1 (Ω, F, P) = ((0, 1), B(0, 1), Lebesgue). be our probability space. Then there exists as sequence R n of independent Bernoulli(1/2) random variables. Proposition. Let (Ω, F, P) = ((0, 1), B(0, 1), Lebesgue). Given any sequence (F n ) of distribution functions, there is a sequence (X n ) of independent random variables with F Xn = F n for all n. 2.4 Convergence of measurable functions Theorem. (i) If µ(e) <, then f n f a.e. implies f n f in measure. (ii) For any E, if f n f in measure, then there exists a subsequence (f nk ) such that f nk f a.e. Theorem (Skorokhod representation theorem of weak convergence). (i) If (X n ), X are defined on the same probability space, and X n X in probability. Then X n X in distribution. (ii) If X n X in distribution, then there exists random variables ( X n ) and X defined on a common probability space with F Xn = F Xn and F X = F X such that X n X a.s. 6

2 Measurable functions and random variables II Probability and Measure (Theorems) 2.5 Tail events Theorem (Kolmogorov 0-1 law). Let (X n ) be a sequence of independent (realvalued) random variables. If A T, then P[A] = 0 or 1. Moreover, if X is a T -measurable random variable, then there exists a constant c such that P[X = c] = 1. 7

3 Integration II Probability and Measure (Theorems) 3 Integration 3.1 Definition and basic properties Proposition. A function is simple iff it is measurable, non-negative, and takes on only finitely-many values. Proposition. Let f : [0, 1] R be Riemann integrable. Then it is also Lebesgue integrable, and the two integrals agree. Theorem (Monotone convergence theorem). Suppose that (f n ), f are nonnegative measurable with f n f. Then µ(f n ) µ(f). Theorem. Let f, g be non-negative measurable, and α, β 0. We have that (i) µ(αf + βg) = αµ(f) + βµ(g). (ii) f g implies µ(f) µ(g). (iii) f = 0 a.e. iff µ(f) = 0. Theorem. Let f, g be integrable, and α, β 0. We have that (i) µ(αf + βg) = αµ(f) + βµ(g). (ii) f g implies µ(f) µ(g). (iii) f = 0 a.e. implies µ(f) = 0. Proposition. If A is a π-system with E A and σ(a) = E, and f is an integrable function that µ(f1 A ) = 0 for all A A. Then µ(f) = 0 a.e. Proposition. Suppose that (g n ) is a sequence of non-negative measurable functions. Then we have ( ) µ g n = µ(g n ). n=1 3.2 Integrals and limits Theorem (Fatou s lemma). Let (f n ) be a sequence of non-negative measurable functions. Then µ(lim inf f n ) lim inf µ(f n ). Theorem (Dominated convergence theorem). Let (f n ), f be measurable with f n (x) f(x) for all x E. Suppose that there is an integrable function g such that f n g for all n, then we have as n. n=1 µ(f n ) µ(f) 8

3 Integration II Probability and Measure (Theorems) 3.3 New measures from old Lemma. For (E, E, µ) a measure space and A E, the restriction to A is a measure space. Proposition. Let (E, E, µ) and (F, F, µ ) be measure spaces and A E. Let f : E F be a measurable function. Then f A is E A -measurable. Proposition. If f is integrable, then f A is µ A -integrable and µ A (f A ) = µ(f1 A ). Proposition. If g is a non-negative measurable function on G, then ν(g) = µ(g f). Proposition. The ν defined above is indeed a measure. 3.4 Integration and differentiation Proposition (Change of variables formula). Let φ : [a, b] R be continuously differentiable and increasing. Then for any bounded Borel function g, we have φ(b) g(y) dy = b φ(a) a g(φ(x))φ (x) dx. ( ) Theorem (Differentiation under the integral sign). Let (E, E, µ) be a space, and U R be an open set, and f : U E R. We assume that (i) For any t U fixed, the map x f(t, x) is integrable; (ii) For any x E fixed, the map t f(t, x) is differentiable; (iii) There exists an integrable function g such that f (t, x) t g(x) for all x E and t U. Then the map x f (t, x) t is integrable for all t, and also the function F (t) = f(t, x)dµ is differentiable, and F (t) = E E f (t, x) dµ. t 9

3 Integration II Probability and Measure (Theorems) 3.5 Product measures and Fubini s theorem Lemma. Let E = E 1 E 2 be a product of σ-algebras. Suppose f : E R is E-measurable function. Then (i) For each x 2 E 2, the function x 1 f(x 1, x 2 ) is E 1 -measurable. (ii) If f is bounded or non-negative measurable, then f 2 (x 2 ) = f(x 1, x 2 ) µ 1 (dx 1 ) E 1 is E 2 -measurable. Theorem. There exists a unique measurable function µ = µ 1 µ 2 on E such that µ(a 1 A 2 ) = µ(a 1 )µ(a 2 ) for all A 1 A 2 A. Theorem (Fubini s theorem). (i) If f is non-negative measurable, then ( ) µ(f) = f(x 1, x 2 ) µ 2 (dx 2 ) µ 1 (dx 1 ). ( ) E 2 E 1 E 1 In particular, we have ( ) f(x 1, x 2 ) µ 2 (dx 2 ) µ 1 (dx 1 ) = E 2 This is sometimes known as Tonelli s theorem. E 2 ( ) f(x 1, x 2 ) µ 1 (dx 1 ) µ 2 (dx 2 ). E 1 (ii) If f is integrable, and { } A = x 1 E : f(x 1, x 2 ) µ 2 (dx 2 ) < E 2. then If we set µ 1 (E 1 \ A) = 0. { f 1 (x 1 ) = E 2 f(x 1, x 2 ) µ 2 (dx 2 ) x 1 A 0 x 1 A, then f 1 is a µ 1 integrable function and µ 1 (f 1 ) = µ(f). Proposition. Let X 1,, X n be random variables on (Ω, F, P) with values in (E 1, E 1 ),, (E n, E n ) respectively. We define E = E 1 E n, E = E 1 E n. Then X = (X 1,, X n ) is E-measurable and the following are equivalent: 10

3 Integration II Probability and Measure (Theorems) (i) X 1,, X n are independent. (ii) µ X = µ X1 µ Xn. (iii) For any f 1,, f n bounded and measurable, we have [ n ] n E f k (X k ) = E[f k (X k )]. k=1 k=1 11

4 Inequalities and L p spaces II Probability and Measure (Theorems) 4 Inequalities and L p spaces 4.1 Four inequalities Proposition (Chebyshev s/markov s inequality). Let f be non-negative measurable and λ > 0. Then µ({f λ}) 1 λ µ(f). Proposition (Jensen s inequality). Let X be an integrable random variable with values in I. If c : I R is convex, then we have E[c(X)] c(e[x]). Lemma. If c : I R is a convex function and m is in the interior of I, then there exists real numbers a, b such that for all x I, with equality at x = m. c(x) ax + b φ ax + b m Proposition (Hölder s inequality). Let p, q (1, ) be conjugate. Then for f, g measurable, we have µ( fg ) = fg 1 f p g q. When p = q = 2, then this is the Cauchy-Schwarz inequality. Lemma. Let a, b 0 and p 1. Then (a + b) p 2 p (a p + b p ). Theorem (Minkowski inequality). Let p [1, ] and f, g measurable. Then 4.2 L p spaces f + g p f p + g p. Theorem. Let 1 p. Then L p is a Banach space. In other words, if (f n ) is a sequence in L p, with the property that f n f m p 0 as n, m, then there is some f L p such that f n f p 0 as n. 12

4 Inequalities and L p spaces II Probability and Measure (Theorems) 4.3 Orthogonal projection in L 2 Theorem. Let V be a closed subspace of L 2. Then each f L 2 has an orthogonal decomposition f = u + v, where v V and u V. Moreover, for all g V with equality iff g v. Lemma (Pythagoras identity). Lemma (Parallelogram law). f v 2 f g 2 f + g 2 = f 2 + g 2 + 2 f, g. f + g 2 + f g 2 = 2( f 2 + g 2 ). Proposition. The conditional expectation of X given G is the projection of X onto the subspace L 2 (G, P) of G-measurable L 2 random variables in the ambient space L 2 (P). 4.4 Convergence in L 1 (P) and uniform integrability Theorem (Bounded convegence theorem). Suppose X, (X n ) are random variables. Assume that there exists a (non-random) constant C > 0 such that X n C. If X n X in probability, then X n X in L 1. Proposition. Finite unions of uniformly integrable sets are uniformly integrable. Proposition. Let X be an L p -bounded family for some p > 1. Then X is uniformly integrable. Lemma. Let X be a family of random variables. Then X is uniformly integrable if and only if sup{e[ X 1 X >k ] : X X } 0 as k. Corollary. Let X = {X}, where X L 1 (P). Then X is uniformly integrable. Hence, a finite collection of L 1 functions is uniformly integrable. Theorem. Let X, (X n ) be random variables. Then the following are equivalent: (i) X n, X L 1 for all n and X n X in L 1. (ii) {X n } is uniformly integrable and X n X in probability. 13

5 Fourier transform II Probability and Measure (Theorems) 5 Fourier transform 5.1 The Fourier transform Proposition. ˆf f 1, ˆµ µ(r d ). Proposition. The functions ˆf, ˆµ are continuous. 5.2 Convolutions Proposition. For any f L p and ν a probability measure, we have f ν p f p. Proposition. f ν(u) = ˆf(u)ˆν(u). Proposition. Let µ, ν be probability measures, and X, Y be independent variables with laws µ, ν respectively. Then µ ν(u) = ˆµ(u)ˆν(u). 5.3 Fourier inversion formula Theorem (Fourier inversion formula). Let f, ˆf L 1. Then f(x) = 1 ˆf(u)e i(u,x) (2π) d du a.e. Proposition. Let Z N(0, 1). Then φ Z (a) = e u2 /2. Proposition. Let Z = (Z 1,, Z d ) with Z j N(0, 1) independent. Then tz has density 1 g t (x) = 2 e x /(2t). (2πt) d/2 with ĝ t (u) = e u 2 t/2. Lemma. The Fourier inversion formula holds for the Gaussian density function. Proposition. Proposition. Proposition. and f g t 1 f 1. f g t (2πt) d/2 f 1. f g t 1 = ˆfĝ t 1 (2π) d/2 t d/2 ˆf 1, f g t ˆf. 14

5 Fourier transform II Probability and Measure (Theorems) Lemma. The Fourier inversion formula holds for Gaussian convolutions. Theorem (Fourier inversion formula). Let f L 1 and f t (x) = (2π) d ˆf(u)e u 2 t/2 e i(u,x) du = (2π) d f g t (u)e i(u,x) du. Then f t f 1 0, as t 0, and the Fourier inversion holds whenever f, ˆf L 1. Lemma. Suppose that f L p with p [1, ). Then f g t f p 0 as t 0. 5.4 Fourier transform in L 2 Theorem (Plancherel identity). For any function f L 1 L 2, the Plancherel identity holds: ˆf 2 = (2π) d/2 f 2. Theorem. There exists a unique Hilbert space automorphism F : L 2 L 2 such that F ([f]) = [(2π) d/2 ˆf] whenever f L 1 L 2. Here [f] denotes the equivalence class of f in L 2, and we say F : L 2 L 2 is a Hilbert space automorphism if it is a linear bijection that preserves the inner product. 5.5 Properties of characteristic functions Theorem. The characteristic function φ X of a distribution µ X of a random variable X determines µ X. In other words, if X and X are random variables and φ X = φ X, then µ X = µ X Theorem. If φ X is integrable, then µ X has a bounded, continuous density function f X (x) = (2π) d φ X (u)e i(u,x) du. Theorem. Let X, (X n ) be random variables with values in R d. If φ Xn (u) φ X (u) for each u R d, then µ Xn µ X weakly. 5.6 Gaussian random variables Proposition. Let X N(µ, σ 2 ). Then Also, for any a, b R, we have E[X] = µ, var(x) = σ 2. ax + b N(aµ + b, a 2 σ 2 ). Lastly, we have φ X (u) = e iµu u2 σ 2 /2. 15

5 Fourier transform II Probability and Measure (Theorems) Theorem. Let X be Gaussian on R n, and le ta be an m n matrix and b R m. Then (i) AX + b is Gaussian on R m. (ii) X L 2 and its law µ X is determined by µ = E[X] and V = var(x), the covariance matrix. (iii) We have φ X (u) = e i(u,µ) (u,v u)/2. (iv) If V is invertible, then X has a density of ( f X (x) = (2π) n/2 (det V ) 1/2 exp 1 ) 2 (x µ, V 1 (x µ)). (v) If X = (X 1, X 2 ) where X i R ni, then cov(x 1, X 2 ) = 0 iff X 1 and X 2 are independent. 16

6 Ergodic theory II Probability and Measure (Theorems) 6 Ergodic theory Proposition. If f is integrable and Θ is measure-preserving. Then f Θ is integrable and f Θdµ = fdµ. Proposition. If Θ is ergodic and f is invariant, then there exists a constant c such that f = c a.e. Theorem. The shift map Θ is an ergodic, measure preserving transformation. 6.1 Ergodic theorems Lemma (Maximal ergodic lemma). Let f be integrable, and where S 0 (f) = 0 by convention. Then E S = sup S n (f) 0, n 0 {S >0} f dµ 0. Theorem (Birkhoff s ergodic theorem). Let (E, E, µ) be σ-finite and f be integrable. There exists an invariant function f such that and µ( f ) µ( f ), S n (f) n If Θ is ergodic, then f is a constant. f a.e. Theorem (von Neumann s ergodic theorem). Let (E, E, µ) be a finite measure space. Let p [1, ) and assume that f L p. Then there is some function f L p such that S n (f) n f in L p. 17

7 Big theorems II Probability and Measure (Theorems) 7 Big theorems 7.1 The strong law of large numbers Theorem (Strong law of large numbers assuming finite fourth moments). Let (X n ) be a sequence of independent random variables such that there exists µ R and M > 0 such that E[X n ] = µ, E[Xn] 4 M for all n. With S n = X 1 + + X n, we have that S n n µ a.s. as n. Theorem (Strong law of large numbers). Let (Y n ) be an iid sequence of integrable random variables with mean ν. With S n = Y 1 + + Y n, we have 7.2 Central limit theorem S n n ν a.s. Theorem. Let (X n ) be a sequence of iid random variables with E[X i ] = 0 and E[X 2 1 ] = 1. Then if we set then for all x R, we have as n. S n = X 1 + + X n, [ ] x Sn e y2 /2 P x dy = P[N(0, 1) x] n 2π 18