Coin tossing space. 0,1 consisting of all sequences (t n ) n N, represents the set of possible outcomes of tossing a coin infinitely many times.

Similar documents
Compatible probability measures

1.1. MEASURES AND INTEGRALS

Measure and Integration: Concepts, Examples and Exercises. INDER K. RANA Indian Institute of Technology Bombay India

2 Measure Theory. 2.1 Measures

Chapter 4. Measure Theory. 1. Measure Spaces

The Lebesgue Integral

Real Analysis Chapter 1 Solutions Jonathan Conder

Construction of a general measure structure

18.175: Lecture 2 Extension theorems, random variables, distributions

REAL ANALYSIS LECTURE NOTES: 1.4 OUTER MEASURE

Problem set 1, Real Analysis I, Spring, 2015.

Chapter 1. Measure Spaces. 1.1 Algebras and σ algebras of sets Notation and preliminaries

Measures and Measure Spaces

Spring 2014 Advanced Probability Overview. Lecture Notes Set 1: Course Overview, σ-fields, and Measures

Lecture 9: Conditional Probability and Independence

Abstract Measure Theory

Continuum Probability and Sets of Measure Zero

(1) Consider the space S consisting of all continuous real-valued functions on the closed interval [0, 1]. For f, g S, define

13. Examples of measure-preserving tranformations: rotations of a torus, the doubling map

Measures. 1 Introduction. These preliminary lecture notes are partly based on textbooks by Athreya and Lahiri, Capinski and Kopp, and Folland.

In N we can do addition, but in order to do subtraction we need to extend N to the integers

6 CARDINALITY OF SETS

Basic Measure and Integration Theory. Michael L. Carroll

Statistical Inference

MORE ON CONTINUOUS FUNCTIONS AND SETS

DRAFT MAA6616 COURSE NOTES FALL 2015

What is proof? Lesson 1

FUNDAMENTALS OF REAL ANALYSIS by. II.1. Prelude. Recall that the Riemann integral of a real-valued function f on an interval [a, b] is defined as

Real Variables: Solutions to Homework 3

ABSTRACT INTEGRATION CHAPTER ONE

In N we can do addition, but in order to do subtraction we need to extend N to the integers

Indeed, if we want m to be compatible with taking limits, it should be countably additive, meaning that ( )

µ (X) := inf l(i k ) where X k=1 I k, I k an open interval Notice that is a map from subsets of R to non-negative number together with infinity

Measurable functions are approximately nice, even if look terrible.

Lebesgue measure on R is just one of many important measures in mathematics. In these notes we introduce the general framework for measures.

Integration on Measure Spaces

RS Chapter 1 Random Variables 6/5/2017. Chapter 1. Probability Theory: Introduction

Lecture 6 Basic Probability

MATH41011/MATH61011: FOURIER SERIES AND LEBESGUE INTEGRATION. Extra Reading Material for Level 4 and Level 6

We are going to discuss what it means for a sequence to converge in three stages: First, we define what it means for a sequence to converge to zero

Measures. Chapter Some prerequisites. 1.2 Introduction

Lecture 3: Probability Measures - 2

HW 4 SOLUTIONS. , x + x x 1 ) 2

MATHS 730 FC Lecture Notes March 5, Introduction

ADVANCED CALCULUS - MTH433 LECTURE 4 - FINITE AND INFINITE SETS

Lecture 6: The Pigeonhole Principle and Probability Spaces

Lebesgue Measure on R n

Lecture 5 Theorems of Fubini-Tonelli and Radon-Nikodym

Part V. 17 Introduction: What are measures and why measurable sets. Lebesgue Integration Theory

Section 2: Classes of Sets

Lebesgue Measure. Dung Le 1

6 Cosets & Factor Groups

Introduction and Preliminaries

An Introduction to Non-Standard Analysis and its Applications

MATH5011 Real Analysis I. Exercise 1 Suggested Solution

STAT 7032 Probability Spring Wlodek Bryc

Some Basic Notations Of Set Theory

Sequence convergence, the weak T-axioms, and first countability

Basic counting techniques. Periklis A. Papakonstantinou Rutgers Business School

Math /Foundations of Algebra/Fall 2017 Foundations of the Foundations: Proofs

5 Set Operations, Functions, and Counting

1/12/05: sec 3.1 and my article: How good is the Lebesgue measure?, Math. Intelligencer 11(2) (1989),

II - REAL ANALYSIS. This property gives us a way to extend the notion of content to finite unions of rectangles: we define

Measure Theoretic Probability. P.J.C. Spreij

Countability. 1 Motivation. 2 Counting

Chapter 2 Metric Spaces

Lebesgue Measure on R n

Handout 2 (Correction of Handout 1 plus continued discussion/hw) Comments and Homework in Chapter 1

Seunghee Ye Ma 8: Week 2 Oct 6

Measure Theoretic Probability. P.J.C. Spreij

Introduction to Proofs in Analysis. updated December 5, By Edoh Y. Amiran Following the outline of notes by Donald Chalice INTRODUCTION

The main results about probability measures are the following two facts:

3 The language of proof

Lebesgue Measure and The Cantor Set

Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension. n=1

consists of two disjoint copies of X n, each scaled down by 1,

The Caratheodory Construction of Measures

Introduction to Ergodic Theory

2 Construction & Extension of Measures

a + b = b + a and a b = b a. (a + b) + c = a + (b + c) and (a b) c = a (b c). a (b + c) = a b + a c and (a + b) c = a c + b c.

Connectedness. Proposition 2.2. The following are equivalent for a topological space (X, T ).

Properties of the Integers

Analysis Comprehensive Exam Questions Fall 2008

Measure and integration

MAT1000 ASSIGNMENT 1. a k 3 k. x =

Statistics for Financial Engineering Session 2: Basic Set Theory March 19 th, 2006

Real Analysis Notes. Thomas Goller

Math 564 Homework 1. Solutions.

Math 421, Homework #7 Solutions. We can then us the triangle inequality to find for k N that (x k + y k ) (L + M) = (x k L) + (y k M) x k L + y k M

Building Infinite Processes from Finite-Dimensional Distributions

Notes on Measure, Probability and Stochastic Processes. João Lopes Dias

Math 172 HW 1 Solutions

Measure Theory. John K. Hunter. Department of Mathematics, University of California at Davis

The strictly 1/2-stable example

REAL AND COMPLEX ANALYSIS

A Short Review of Cardinality

Introduction to Real Analysis

Nets and filters (are better than sequences)

RANDOM WALKS AND THE PROBABILITY OF RETURNING HOME

Exercises for Unit VI (Infinite constructions in set theory)

Transcription:

Coin tossing space Think of a coin toss as a random choice from the two element set }. Thus the set } n represents the set of possible outcomes of n coin tosses, and Ω := } N, consisting of all sequences (t n ) n N, represents the set of possible outcomes of tossing a coin infinitely many times. If the coin is unbiased and the tosses are independent, each of the 2 n outcomes in } n is equally probable. So } n is equipped with a probability measure µn, which is 2 n times counting measure on } n. (All subsets of } n are measurable.)

Compatible probability measures Assume that m < n. Tossing a coin n times, and then ignoring all but the first m of them, is represented by a projection map π mn : } n } m: π mn (t 1,..., t n ) = (t 1,..., t m ). An event depending on (t 1,..., t m ) is a set E } m. It has probability µm (E). The same event considered as a subset of } n is π 1 mn (E). It does in fact have the same probability, in } n } m:, as E does, in µ n ( π 1 mn (E)) = µ m (E), E } m. (1) Think of this as a compatibility result: All the different probability measures µ n give the same result for those events they can measure. (Exercise: Show this. Hint: Show that for each s E there are 2 n m distinct sequences t π 1 mn (E) with π mn(t) = s.)

Lots and lots of projections All these projection maps π nm are compatible in the sense that π km π mn = π kn, k < m < n. It is helpful to picture the finite coin tossing spaces as an infinite sequence: } } 2 } 3 } n } n+1 where the arrows shown are π 12, π 23,..., π n,n+1. At the far right we can put the infinite coin tossing space Ω. (Category theorists call that the projective limit of the above diagram, but never mind that.) There is also a projection map π n : Ω } n, given by πn (t 1, t 2,...) = (t 1,..., t n ). Clearly, the compatibility of projections extends to this one: π mn π n = π m, m < n. (2)

Finitely determined events and their probabilities Some events in Ω depend only on a finite number of coin tosses. (By this I mean that to determine whether t Ω belongs to some event E, you only need to look at t 1,..., t n for some n, and that n is independent of t.) If E } n then π 1 n (E) = t Ω: (t 1,..., t n ) E } is such an event, and I expect you can convince yourself that every such finitely determined event has this form. (If not, consider that a definition.) Define A n = π 1 n (E): E } n }, n N. Then A n is a σ-algebra of subsets of Ω for each n N, and moreover A m A n when m < n (for the equality π mn π n = π m implies π 1 m (E) = π 1 n (π 1 nm (E)) when E } m ). Let C = A n. n N It is not hard to show that, since (A n ) is an increasing sequence of algebras, the union C is also an algebra of sets (do it!). We can define a function ι on C by ι ( π 1 n (E)) = µ n (E), E } n. (Exercise: Show that this is well defined. Hint: Use (1) and (2).)

Extending the measure The function ι represents the probability of any finitely determined event. But what about other events? How can we discuss the probability that 1 n lim t n n k = 1 2, for example? The set of t Ω satisfying this equality is clearly not finitely determined. What is needed is to extend ι to a measure. It is defined on the algebra C, which is not a σ-algebra. But C is a semialgebra, so in order to show that we can extend ι to a measure, we need to show that ι( ) = 0, that ι is finitely additive, and the appropriate subadditivity condition. For the first property, ι( ) = µ n (π 1 n ( )) = µ n( ) = 0. For the second, if F 1,...,F m C are pairwise disjoint then there is some n so that F k A n for k = 1,...,m. Write F k = π 1 n (E k). Then the sets E k must be pairwise disjoint, and ( m ) ( m ) m m ι F k = µ n E k = µ n (E k ) = ι(f k ). Finally, assume that F,F k C with F k N F k. I claim that a finite number of the sets F k will cover F, i.e., F F 1 F n for some n. If this is true, then ( n ) ι(f ) ι F k n ι(f k ) ι(f k ), which is the subadditivity condition we need to show. To prove the claim, assume instead it is false. That is, we assume that each set A n, defined by A n = F \ (F 1 F n ), is nonempty. Note that A n A, and A 1 A 2 A 3. We now define a sequence (u k ) in } by induction. First, since Ω = π 1 } } 1 (0) π 1 1 (1), we find An = ( A n π 1 }) ( 1 (0) An π 1 }) 1 (1), so we can select u 1 } so that A n π 1 1 (u1 ) } for infinitely many n, and therefore for all n N, since (A n ) is a decreasing sequence of sets. Next, we get A n π 1 1 (u1 ) } = ( A n π 1 2 (u1,0) }) ( A n π 1 1 (u1,1) }), so we can select u 2 } so that A n π 1 1 (u1,u 2 ) } for infinitely many n, and therefore for all n N. Continuing this procedure, we end up with a sequence u = (u k ) k N with the property that, for each k, A n π 1 (u1,...,u k k ) } for all n N. I claim that u n N A n. To see this, merely note that A n (for any n) is finititely determined, i.e., there is some k N so that A n = π 1 k E with E } k. But since A n π 1 (u1,...,u k k ) }, it follows that (u 1,...,u k ) E, and therefore u A n. We have proved that n N A n. This contradicts the assumption that F k N F k, and the claim is therefore proved.

Probability on the coin tossing space From the general theory, we now have a measure µ on a σ-algebra A extending C, and hence all the algebras A n. Moreover, by construction µ ( π 1 n (E)) = µ n (E) for all E } n, so our probability measure certainly captures our intuition about the distribution of n coin tosses. The nth toss is represented by a function T n : Ω }, defined by T n (t) = t n. T n is a stochastic variable, i.e., a measurable function on Ω. It is not difficult to show that the tosses T n are stochastically independent: If G 1,...,G n R are Borel sets, then ( n ) µ T 1 k (G k) = n µ ( T 1 k (G k) ). (Note that T 1 k (G k) is one of four sets:, t Ω: t k = 0 }, t Ω: t k = 1 }, or Ω.)

Interlude: Pulling and pushing measures If (Λ,A,µ) is a measure space and g : Ω Λ is a given function, (where the doubleheaded arrow indicates that the map is onto), we can define the pullback σ-algebra g A and measure g µ on Ω: g A = g 1 (A): A A }, g µ ( g 1 (A) ) = µ(a) for A A. The proof is an easy exercise. (The assumption that g is onto is important.) Note that we used this construction above, using the projection map π n to pull the probablitiy measure µ n back from } n to Ω. We can also push a measure forward: Given a on map g : Ω Λ where (Ω,A,µ) is a measure space, we can create the pushforward σ-algebra g A and measure g µ on Λ: Again, the proof is an easy exercise. g A = A Λ: g 1 (A) A }, g µ(a) = µ ( g 1 (A) ) for A g A. A typical example of this is the distribution of a real-valued stochastic variable X : Ω R on a probability space (Ω,A,µ): Since X is by definition measurable, any interval, and therefore any Borel set, belongs to X A. The restriction of the pushforward measure X µ to the σ-algebra B of Borel sets is called the distribution of X. In particular, the cumulative distribution function F of X is given by F (x) = µ ( X x }) = µ ( X 1 ((, x]) ) = X µ((, x]) and so, the distribution of X turns out to be the Lebesgue Stieltjes measure associated with F. Warning: I skipped quite a few details in this section. The sections to follow will skip even more details. You are encouraged to try to fill in the holes yourself.

Pushing coin tossing measure We can map the coin tossing space Ω onto [] via the map f : Ω [] given by f (t) = t k 2 k, k N which boils down to interpreting the coin tossing sequence t = (t) n N as a sequence of binary digits for the number f (t). Since the kth toss map t t k is clearly a measurable function, then so is f. Let us investigate the pushforward measure f µ on []. A real number is called a dyadic rational number if it can be written as 2 n m for some integers m and n. A moment s reflection shows that, given a dyadic rational x = 2 n m [), the inverse image f 1( [x, x + 2 n ] ) is given by those t Ω whose first n components correspond to the first n bits of x, plus one more sequence ending in infinitely many 1s. For example, f 1 ([ 3 8, 4 8 ]) = t Ω: (t 1, t 2, t 3 ) = (,1) } (,,1,1,1,...) }. Since any singleton set in Ω has measure zero, it follows (not immediately, but with a modest amount of work) that f µ ( [x, y] ) = y x = λ ( [x, y] ) for any dyadic rational x, y in [] with x y, with the same result for (x, y), [x, y) and (x, y]. Since such intervals form a semialgebra which generates the Borel sets, the uniqueness theorem for the extension of measures allows us to conclude the following: The pushforward f µ equals Lebesgue measure on []. Strictly speaking, the uniqueness theorem only allows us to conclude that f µ(e) = λ(e) for Borel sets contained in [], but this is easily extended to Lebesgue measurable sets, since both µ and λ are complete measures.

Pushing coin tossing measure some more We are not done pushing coin tossing measure to [0, 1]. g : Ω C given by g (t) = 2t k 3 k, k N Consider the function noting that this maps Ω onto the Cantor set C []. This map is easy to understand: Take for example t = (,1,,1,...): Since t 1 = 0, g (t) belongs to the left interval of length 1 3 left after we removed the middle of []. Next, t 2 = 1, so g (t) belong to the right interval of length 1 9 left over after we removed the middle of [0, 1 3 ] (and [ 2 3,1]). Next, g (t) belogs to the right interval of length 1 27 left over after we removed the middle of [ 2 9, 3 9 ] (and three other intervals). It should be clear that g is one to one, for as soon as we have encountered a difference between s and t, we know that g (s) and g (t) belongs to well separated intervals. The resulting measure g µ on C is sometimes referred to as the Cantor measure. Its cumulative distribution function is none other than the standard Cantor function ψ. You may verify this directly, but noting that both functions take the same constant value on each of the open intervals in [] \ C. Another approach is to prove and then use the relation ψ ( g (t) ) = f (t), where f is the function defined in the previous paragraph.

Infinitely many uniform variables We return to the map f : Ω []. We can think of this map as a stochastic variable uniformly distributed on []. In probability theory, one often needs infinitely many stochastically independent such variables. And cointossing space delivers! The easiest way to do this is to notice that N 2 = N N is countable, so there is a one-to-one map c : N 2 N. We can use it to define U n : Ω []: U n (t) = t c(n,k) 2 k, n N. k N There are straightforward techniques for transforming a uniformly distributed variable into a stochastic variable with any desired distribution, so you can easily generate infinitely many independent variables with a standard normal distribution, for example. And that gives you the building blocks to generate a model for Brownian motion (also known as the Wiener process). But that is a topic for a different course.