Average laws in analysis

Similar documents
Three hours THE UNIVERSITY OF MANCHESTER. 31st May :00 17:00

Probability and Measure

Dynamics and time series: theory and applications. Stefano Marmi Giulio Tiozzo Scuola Normale Superiore Lecture 2, Jan 14, 2010

An Introduction to Ergodic Theory

DYNAMICAL SYSTEMS PROBLEMS. asgor/ (1) Which of the following maps are topologically transitive (minimal,

Rudiments of Ergodic Theory

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

Chapter 3. Regular grammars

Waiting times, recurrence times, ergodicity and quasiperiodic dynamics

Lyapunov Exponents of Linear Cocycles *

Ergodic Theory and Topological Groups

1 Sequences of events and their limits

25.1 Ergodicity and Metric Transitivity

PHY411 Lecture notes Part 5

Ergodic Theory. Constantine Caramanis. May 6, 1999

An Introduction to Ergodic Theory Normal Numbers: We Can t See Them, But They re Everywhere!

ACO Comprehensive Exam 19 March Graph Theory

Dynamical Systems and Ergodic Theory PhD Exam Spring Topics: Topological Dynamics Definitions and basic results about the following for maps and

Concentration inequalities for linear cocycles and their applications to problems in dynamics and mathematical physics *

String Matching. Thanks to Piotr Indyk. String Matching. Simple Algorithm. for s 0 to n-m. Match 0. for j 1 to m if T[s+j] P[j] then

4. Ergodicity and mixing

Algebra 2 CP Final Exam Review

DETERMINISTIC REPRESENTATION FOR POSITION DEPENDENT RANDOM MAPS

Dynamical Systems and Ergodic Theory

arxiv: v1 [math.ds] 1 Oct 2014

5 Birkhoff s Ergodic Theorem

University of York. Extremality and dynamically defined measures. David Simmons. Diophantine preliminaries. First results. Main results.

Roth s Theorem on Arithmetic Progressions

EECS 229A Spring 2007 * * (a) By stationarity and the chain rule for entropy, we have

Part I: Definitions and Properties

HILBERT SPACES AND THE RADON-NIKODYM THEOREM. where the bar in the first equation denotes complex conjugation. In either case, for any x V define

THE STRUCTURE OF THE SPACE OF INVARIANT MEASURES

ALMOST SURE CONVERGENCE OF RANDOM GOSSIP ALGORITHMS

ATotallyDisconnectedThread: Some Complicated p-adic Fractals

FURSTENBERG S ERGODIC THEORY PROOF OF SZEMERÉDI S THEOREM

A Study of Probability and Ergodic theory with applications to Dynamical Systems

arxiv: v2 [math.ds] 24 Apr 2018

MAGIC010 Ergodic Theory Lecture Entropy

Lecture Notes Introduction to Ergodic Theory

An Invitation to Modern Number Theory. Steven J. Miller and Ramin Takloo-Bighash PRINCETON UNIVERSITY PRESS PRINCETON AND OXFORD

UNIVERSITY OF BRISTOL. Mock exam paper for examination for the Degrees of B.Sc. and M.Sci. (Level 3)

13. Examples of measure-preserving tranformations: rotations of a torus, the doubling map

MET Workshop: Exercises

Bounds for (generalised) Lyapunov exponents for deterministic and random products of shears

Admin and Lecture 1: Recap of Measure Theory

3 hours UNIVERSITY OF MANCHESTER. 22nd May and. Electronic calculators may be used, provided that they cannot store text.

S-adic sequences A bridge between dynamics, arithmetic, and geometry

Dynamical Systems and Ergodic Theory

DYNAMICAL CUBES AND A CRITERIA FOR SYSTEMS HAVING PRODUCT EXTENSIONS

A Short Introduction to Ergodic Theory of Numbers. Karma Dajani

Rigorous approximation of invariant measures for IFS Joint work April 8, with 2016 S. 1 Galat / 21

ABSTRACT CONDITIONAL EXPECTATION IN L 2

Introduction: Pythagorean Triplets

Lecture 10. Theorem 1.1 [Ergodicity and extremality] A probability measure µ on (Ω, F) is ergodic for T if and only if it is an extremal point in M.

ERGODIC AVERAGES FOR INDEPENDENT POLYNOMIALS AND APPLICATIONS

CHILE LECTURES: MULTIPLICATIVE ERGODIC THEOREM, FRIENDS AND APPLICATIONS

Bernoulli decompositions and applications

Pi: The Ultimate Ratio

Dept of Math., SCU+USTC

INITIAL POWERS OF STURMIAN SEQUENCES

Lecture 2: Convergence of Random Variables

Approximate Transitivity for Zero Entropy Systems

Lecture 7. Sums of random variables

MATH41112/ Ergodic Theory. Charles Walkden

Computational Models #1

EE376A: Homework #2 Solutions Due by 11:59pm Thursday, February 1st, 2018

Szemerédi s Theorem via Ergodic Theory

Zeros of lacunary random polynomials

Periodic Sinks and Observable Chaos

Part II Probability and Measure

Quantitative recurrence for beta expansion. Wang BaoWei

Academic Outcomes Mathematics

Computer Sciences Department

New rotation sets in a family of toral homeomorphisms

A REMARK ON SLUTSKY S THEOREM. Freddy Delbaen Departement für Mathematik, ETH Zürich

FURSTENBERG S THEOREM ON PRODUCTS OF I.I.D. 2 ˆ 2 MATRICES

Lectures 6 7 : Marchenko-Pastur Law

arxiv: v2 [math.ds] 14 Apr 2011

Convergence of Random Variables

Continued Fraction Digit Averages and Maclaurin s Inequalities

STA205 Probability: Week 8 R. Wolpert

Entropy and Ergodic Theory Notes 21: The entropy rate of a stationary process

Self-similar Fractals: Projections, Sections and Percolation

Theorem 2.1 (Caratheodory). A (countably additive) probability measure on a field has an extension. n=1

Propp-Wilson Algorithm (and sampling the Ising model)

On Equivariant Block Triangularization of Matrix Cocyles

18.175: Lecture 3 Integration

A dyadic endomorphism which is Bernoulli but not standard

Module 1: Analyzing the Efficiency of Algorithms

Lecture 15: Conditional and Joint Typicaility

Check boxes of Edited Copy of Sp Topics (was 261-pilot)

Ron Paul Curriculum Mathematics 8 Lesson List

IRRATIONAL ROTATION OF THE CIRCLE AND THE BINARY ODOMETER ARE FINITARILY ORBIT EQUIVALENT

Probability and Measure

consists of two disjoint copies of X n, each scaled down by 1,

An Introduction to Entropy and Subshifts of. Finite Type

The recurrence time for interval exchange maps

Countability Sets of Measure Zero Random Reals Normal Numbers There s a Bear in There. The Real Thing. Paul McCann. Wednesday, 3 August, 2011

Discrete Mathematics. Benny George K. September 22, 2011

Dynamics and time series: theory and applications. Stefano Marmi Giulio Tiozzo Scuola Normale Superiore Lecture 3, Jan 20, 2010

Transcription:

Average laws in analysis Silvius Klein Norwegian University of Science and Technology (NTNU)

The law of large numbers: informal statement The theoretical expected value of an experiment is approximated by the average of a large number of independent samples. theoretical expected value empirical average

The law of large numbers (LLN) Let X 1, X 2,..., X n,... be a sequence of jointly independent, identically distributed copies of a scalar random variable X. Assume that X is absolutely integrable, with expectation µ. Define the partial sum process Then the average process S n := X 1 + X 2 +... + X n. S n n µ as n.

The law of large numbers: formal statements Let X 1, X 2,... be a sequence of independent, identically distributed random variables with common expectation µ. Let S n := X 1 + X 2 +... + X n be the corresponding partial sums process. Then 1 (weak LLN) S n n That is, for every ɛ > 0, µ in probability. P { S n n µ > ɛ } 0 as n. 2 (strong LLN) S n n µ almost surely.

It was the best of times, it was the worst of times. Charles Dickens, A tale of two cities (click here)

Application of LLN: the infinite monkey theorem Let X 1, X 2,... be i.i.d. random variables drawn uniformly from a finite alphabet. Then almost surely, every finite phrase (i.e. finite string of symbols in the alphabet) appears (infinitely often) in the string X 1 X 2 X 3....

Application of LLN: the infinite monkey theorem Let X 1, X 2,... be i.i.d. random variables drawn uniformly from a finite alphabet. Then almost surely, every finite phrase (i.e. finite string of symbols in the alphabet) appears (infinitely often) in the string X 1 X 2 X 3.... yskpw,qol,all/alkmas;.a ma;;lal;,qwmswl,;q;[; lkle 78623rhbkbads m,q l;, ;f.w, fwe It was the best of times, it was the worst of times. jllkasjllmk,a s.,qjwejhns;.2;oi0ppk;q,qkjkqhjnqnmnmmasi[oqw qqnkm,sa;l;[ml/w/ q

Application of LLN: the infinite monkey theorem Let X 1, X 2,... be i.i.d. random variables drawn uniformly from a finite alphabet. Then almost surely, every finite phrase (i.e. finite string of symbols in the alphabet) appears (infinitely often) in the string X 1 X 2 X 3.... yskpw,qol,all/alkmas;.a ma;;lal;,qwmswl,;q;[; lkle 78623rhbkbads m,q l;, ;f.w, fwe It was the best of times, it was the worst of times. jllkasjllmk,a s.,qjwejhns;.2;oi0ppk;q,qkjkqhjnqnmnmmasi[oqw qqnkm,sa;l;[ml/w/ q

The second Borel-Cantelli lemma Let E 1, E 2,..., E n,... be a sequence of jointly independent events. If P(E n ) =, n=1 then almost surely, an infinite number of E n hold simultaneously. This can be deduced from the strong law of large numbers, applied to the random variables X k := 1 Ek.

The actual proof of the infinite monkey theorem Split every realization of the infinite string of symbols in the alphabet X 1 X 2 X 3... X n... into finite strings S 1, S 2,... of length 52 each. Let E n be the event that the phrase It was the best of times, it was the worst of times. is exactly the n-th finite string S n. These are independent events. They each have the same probability p > 0 to occur. Apply the second Borel-Cantelli lemma.

The law of large numbers We have seen that if X 1, X 2,..., X n,... is a sequence of jointly independent, identically distributed copies of a scalar random variable X, and if we denote the corresponding sum process by then the average process S n := X 1 + X 2 +... + X n, S n n E X as n.

A rather deterministic system: circle rotations Let S be the unit circle in the (complex) plane. There is a natural measure λ on S (i.e. the extension of the arc-length). Let 2πα be an angle, and denote by R α the rotation by 2πα on S. That is, consider the transformation R α : S S, where if z = e 2π i x S and if we denote ω := e 2π i α, then R α (z) = e 2π i (x+α) = z ω. Note that R α preserves the measure λ.

Iterations of the circle rotation Let 2πα be an angle. Start with a point z = e 2π i x S and consider successive applications of the rotation map R α : R 1 α(z) = R α (z) R 2 α(z) = R α R α (z). 2π i (x+α) = e 2π i (x+2α) = e Rα(z) n 2π i (x+nα) = R α... R α (z) = e. The maps R 1 α, R 2 α,..., R n α,... are the iterations of R α. Given a point z S, the set is called the orbit of z. {R 1 α(z), R 2 α(z),..., R n α(z),... }

An orbit of a circle rotation Let R α be the circle rotation by the angle 2πα, where α is an irrational number. Pick a point z on the circle S. The orbit of z (or rather a finite subset of it).

An orbit of a circle rotation Let R α be the circle rotation by the angle 2πα, where α is an irrational number. Pick a point z on the circle S. The orbit of z (or rather a finite subset of it). The orbit of every point is dense on the circle. This transformation satisfies a very weak form of independence called ergodicity.

Observables on the unit circle Any measurable function f : S R is called a (scalar) observable of the measure space (S, A, λ). We will assume our observables to be absolutely integrable. A basic example of an observable: f = 1 I, where I is an arc (or any other measurable set) on the circle. I I ~ e, Observations" of the orbit points of a circle rotation.

Average number of orbit points visiting an arc Let R α be the circle rotation by the angle 2πα, where α is an irrational number. Let I be an arc on the circle. I I ~ e, The first n orbit points of a circle rotation and their visits to I. The average number of visits to I: { } # j {1, 2,..., n} : Rα(z) j I n What does this look like for large enough n? Or in other words, is there a limit of these averages as n?

Average number of orbit points visiting an arc Let R α be the circle rotation by the angle 2πα, where α is an irrational number. Let I be an arc on the circle. I I ~ e, The first n orbit points of a circle rotation and their visits to I. As n, the average number of visits to I: { } # j {1, 2,..., n} : Rα(z) j I λ(i), n for all points z S.

Average number of orbit points visiting an arc Let R α be the circle rotation by the angle 2πα, where α is an irrational number. Let I be an arc on the circle. I I ~ e, The first n orbit points of a circle rotation and their visits to I. { } # j {1, 2,..., n} : Rα(z) j I = n 1 I (Rα(z)). j j=1 Then the average number of visits to I can be written: 1 I (R 1 α(z)) + 1 I (R 2 α(z)) +... + 1 I (R n α(z)) n λ(i)

Average number of orbit points visiting an arc Let R α be the circle rotation by the angle 2πα, where α is an irrational number. Let I be an arc on the circle. I I ~ e, The first n orbit points of a circle rotation and their visits to I. { } # j {1, 2,..., n} : Rα(z) j I = n 1 I (Rα(z)). j Then the average number of visits to I can be written: 1 I (Rα(z)) 1 + 1 I (Rα(z)) 2 +... + 1 I (Rα(z)) n λ(i) = 1 I dλ. n j=1 S

Measure preserving dynamical systems A probability space (X, B, µ) together with a transformation T : X X define a measure preserving dynamical system if T is measurable and it preserves the measure of any B-measurable set: µ(t 1 A) = µ(a) for all A B. Ergodic dynamical system. For any B-measurable set A with µ(a) > 0, the iterations T A, T 2 A,..., T n A,... fill up the whole space X, except possibly for a set of measure zero. Ergodicity leads to some very, very weak form of independence.

Some examples of ergodic dynamical systems 1 The Bernoulli shift, which encodes sequences of independent, identically distributed random variables. 2 The circle rotation by an irrational angle. 3 The doubling map. T : [0, 1] [0, 1], Tx = 2x mod 1..

The pointwise ergodic theorem Given: an ergodic dynamical system (X, B, µ, T), and an absolutely integrable observable f : X R, define the n-th Birkhoff sum S n f (x) := f (Tx) + f (T 2 x) +... + f (T n x). Then as n 1 n S n f (x) X f dµ for µ a.e. x X.

The law of large numbers We have seen that if X 1, X 2,..., X n,... is a sequence of jointly independent, identically distributed copies of a scalar random variable X, and if we denote the corresponding sum process by then as n S n := X 1 + X 2 +... + X n, 1 n S n X almost surely.

An immediate application of the ergodic theorem Let (X, B, µ, T) be an ergodic dynamical system. Let x X, and consider its orbit Tx, T 2 x,..., T n x,... Equidistribution of orbit points. For any B-measurable set A, the average number of orbit points that visit A, converges as n. { } # j {1, 2,..., n} : T j x A µ(a). n for µ almost every point x X. Proof. Just apply the pointwise ergodic theorem to the observable f = 1 A, and note that the counting of orbit points above equals the n-th Birkhoff sum of this observable.

Another simple application of the ergodic theorem Consider the decimal representation of every real number x [0, 1). x = 0. x 1 x 2... x n..., where the digits x k {0, 1, 2,..., 9}. What is the frequency (average occurrence) of each digit in the decimal representation of a typical" real number x [0, 1]?

Another simple application of the ergodic theorem Consider the decimal representation of every real number x [0, 1). x = 0. x 1 x 2... x n..., where the digits x k {0, 1, 2,..., 9}. What is the frequency (average occurrence) of each digit in the decimal representation of a typical" real number x [0, 1]? { } # j {1, 2,..., n} : x j = 7? as n. n

Another simple application of the ergodic theorem Consider the decimal representation of every real number x [0, 1). x = 0. x 1 x 2... x n..., where the digits x k {0, 1, 2,..., 9}. What is the frequency (average occurrence) of each digit in the decimal representation of a typical" real number x [0, 1]? { } # j {1, 2,..., n} : x j = 7? as n. n Solution. Consider the dynamical system given by the 10-fold map: T : [0, 1) [0, 1), Tx = 10x mod 1. Let f : [0, 1) R be the observable defined as { 1 if x1 = 7 f (x) = 0 otherwise.

The law of large numbers We have seen that if X 1, X 2,..., X n,... is a sequence of jointly independent, identically distributed scalar random variables, and if we denote the corresponding sum process by S n := X 1 + X 2 +... + X n, then the arithmetic averages 1 n S n converge almost surely as n.

Random matrices and geometric averages Consider a sequence M 1, M 2,..., M n,... of random matrices. We assume that this sequence is independent and identically distributed. Consider the partial products process: Π n = M n... M 2 M 1.

Random matrices and geometric averages Consider a sequence M 1, M 2,..., M n,... of random matrices. We assume that this sequence is independent and identically distributed. Consider the partial products process: Π n = M n... M 2 M 1. Furstenberg-Kesten s theorem. Almost surely, and as n, the geometric averages" 1 n log Π n converge to a constant. This constant is called the Lyapunov exponent of the multiplicative process.

If you liked this...

If you liked this...... then MA3105 Advanced real analysis will cover in depth many of these topics.

If you liked this...... then MA3105 Advanced real analysis will cover in depth many of these topics. You are ready to take MA3105 if this picture makes some sense to you.

MA3105 Advanced real analysis main topics A monkey typing random stuff.

MA3105 Advanced real analysis main topics Arnold s cat map.