18.175: Lecture 14 Infinite divisibility and so forth

Size: px
Start display at page:

Download "18.175: Lecture 14 Infinite divisibility and so forth"

Transcription

1 Lecture : Lecture 14 Infinite divisibility and so forth Scott Sheffield MIT

2 Lecture 14 Outline Infinite divisibility Higher dimensional CFs and CLTs Random walks Stopping times Arcsin law, other SRW stories

3 Lecture 14 Outline Infinite divisibility Higher dimensional CFs and CLTs Random walks Stopping times Arcsin law, other SRW stories

4 Infinitely divisible laws Say a random variable X is infinitely divisible, for each n, there is a random variable Y such that X has the same law as the sum of n i.i.d. copies of Y.

5 Infinitely divisible laws Say a random variable X is infinitely divisible, for each n, there is a random variable Y such that X has the same law as the sum of n i.i.d. copies of Y. What random variables are infinitely divisible?

6 Infinitely divisible laws Say a random variable X is infinitely divisible, for each n, there is a random variable Y such that X has the same law as the sum of n i.i.d. copies of Y. What random variables are infinitely divisible? Poisson, Cauchy, normal, stable, etc.

7 Infinitely divisible laws Say a random variable X is infinitely divisible, for each n, there is a random variable Y such that X has the same law as the sum of n i.i.d. copies of Y. What random variables are infinitely divisible? Poisson, Cauchy, normal, stable, etc. Let s look at the characteristic functions of these objects. What about compound Poisson random variables (linear combinations of independent Poisson random variables)? What are their characteristic functions like?

8 Infinitely divisible laws Say a random variable X is infinitely divisible, for each n, there is a random variable Y such that X has the same law as the sum of n i.i.d. copies of Y. What random variables are infinitely divisible? Poisson, Cauchy, normal, stable, etc. Let s look at the characteristic functions of these objects. What about compound Poisson random variables (linear combinations of independent Poisson random variables)? What are their characteristic functions like? What if have a random variable X and then we choose a Poisson random variable N and add up N independent copies of X.

9 Infinitely divisible laws Say a random variable X is infinitely divisible, for each n, there is a random variable Y such that X has the same law as the sum of n i.i.d. copies of Y. What random variables are infinitely divisible? Poisson, Cauchy, normal, stable, etc. Let s look at the characteristic functions of these objects. What about compound Poisson random variables (linear combinations of independent Poisson random variables)? What are their characteristic functions like? What if have a random variable X and then we choose a Poisson random variable N and add up N independent copies of X. More general constructions are possible via Lévy Khintchine representation.

10 Lecture 14 Outline Infinite divisibility Higher dimensional CFs and CLTs Random walks Stopping times Arcsin law, other SRW stories

11 Lecture 14 Outline Infinite divisibility Higher dimensional CFs and CLTs Random walks Stopping times Arcsin law, other SRW stories

12 Higher dimensional limit theorems Much of the CLT story generalizes to higher dimensional random variables.

13 Lecture 14 Higher dimensional limit theorems Much of the CLT story generalizes to higher dimensional random variables. For example, given a random vector (X, Y, Z), we can define φ(a, b, c) = Ee i(ax +by +cz).

14 Lecture 14 Higher dimensional limit theorems Much of the CLT story generalizes to higher dimensional random variables. For example, given a random vector (X, Y, Z), we can define φ(a, b, c) = Ee i(ax +by +cz). This is just a higher dimensional Fourier transform of the density function.

15 Lecture 14 Higher dimensional limit theorems Much of the CLT story generalizes to higher dimensional random variables. For example, given a random vector (X, Y, Z), we can define φ(a, b, c) = Ee i(ax +by +cz). This is just a higher dimensional Fourier transform of the density function. The inversion theorems and continuity theorems that apply here are essentially the same as in the one-dimensional case.

16 Lecture 14 Outline Infinite divisibility Higher dimensional CFs and CLTs Random walks Stopping times Arcsin law, other SRW stories

17 Lecture 14 Outline Infinite divisibility Higher dimensional CFs and CLTs Random walks Stopping times Arcsin law, other SRW stories

18 Lecture 14 Exchangeable events Start with measure space (S, S, µ). Let Ω = {(ω 1, ω 2,...) : ω i S}, let F be product σ-algebra and P the product probability measure.

19 Lecture 14 Exchangeable events Start with measure space (S, S, µ). Let Ω = {(ω 1, ω 2,...) : ω i S}, let F be product σ-algebra and P the product probability measure. Finite permutation of N is one-to-one map from N to itself that fixes all but finitely many points.

20 Exchangeable events Start with measure space (S, S, µ). Let Ω = {(ω 1, ω 2,...) : ω i S}, let F be product σ-algebra and P the product probability measure. Finite permutation of N is one-to-one map from N to itself that fixes all but finitely many points. Event A F is permutable if it is invariant under any finite permutation of the ω i.

21 Exchangeable events Start with measure space (S, S, µ). Let Ω = {(ω 1, ω 2,...) : ω i S}, let F be product σ-algebra and P the product probability measure. Finite permutation of N is one-to-one map from N to itself that fixes all but finitely many points. Event A F is permutable if it is invariant under any finite permutation of the ω i. Let E be the σ-field of permutable events.

22 Exchangeable events Start with measure space (S, S, µ). Let Ω = {(ω 1, ω 2,...) : ω i S}, let F be product σ-algebra and P the product probability measure. Finite permutation of N is one-to-one map from N to itself that fixes all but finitely many points. Event A F is permutable if it is invariant under any finite permutation of the ω i. Let E be the σ-field of permutable events. This is related to the tail σ-algebra we introduced earlier in the course. Bigger or smaller?

23 Hewitt-Savage 0-1 law If X 1, X 2,... are i.i.d. and A A then P(A) {0, 1}.

24 Hewitt-Savage 0-1 law If X 1, X 2,... are i.i.d. and A A then P(A) {0, 1}. Idea of proof: Try to show A is independent of itself, i.e., that P(A) = P(A A) = P(A)P(A). Start with measure theoretic fact that we can approximate A by a set A n in σ-algebra generated by X 1,... X n, so that symmetric difference of A and A n has very small probability. Note that A n is independent of event A n that A n holds when X 1,..., X n and X n1,..., X 2n are swapped. Symmetric difference between A and A n is also small, so A is independent of itself up to this small error. Then make error arbitrarily small.

25 Application of Hewitt-Savage: If X i are i.i.d. in R n then S n = n i=1 X i is a random walk on R n.

26 Application of Hewitt-Savage: If X i are i.i.d. in R n then S n = n i=1 X i is a random walk on R n. Theorem: if S n is a random walk on R then one of the following occurs with probability one:

27 Application of Hewitt-Savage: If X i are i.i.d. in R n then S n = n i=1 X i is a random walk on R n. Theorem: if S n is a random walk on R then one of the following occurs with probability one: Sn = 0 for all n

28 Application of Hewitt-Savage: If X i are i.i.d. in R n then S n = n i=1 X i is a random walk on R n. Theorem: if S n is a random walk on R then one of the following occurs with probability one: Sn = 0 for all n Sn

29 Application of Hewitt-Savage: If X i are i.i.d. in R n then S n = n i=1 X i is a random walk on R n. Theorem: if S n is a random walk on R then one of the following occurs with probability one: Sn = 0 for all n Sn Sn

30 Application of Hewitt-Savage: If X i are i.i.d. in R n then S n = n i=1 X i is a random walk on R n. Theorem: if S n is a random walk on R then one of the following occurs with probability one: Sn = 0 for all n Sn Sn = lim inf S n < lim sup S n =

31 Application of Hewitt-Savage: If X i are i.i.d. in R n then S n = n i=1 X i is a random walk on R n. Theorem: if S n is a random walk on R then one of the following occurs with probability one: Sn = 0 for all n Sn Sn = lim inf S n < lim sup S n = Idea of proof: Hewitt-Savage implies the lim sup S n and lim inf S n are almost sure constants in [, ]. Note that if X 1 is not a.s. constant, then both values would depend on X 1 if they were not in ±

32 Lecture 14 Outline Infinite divisibility Higher dimensional CFs and CLTs Random walks Stopping times Arcsin law, other SRW stories

33 Lecture 14 Outline Infinite divisibility Higher dimensional CFs and CLTs Random walks Stopping times Arcsin law, other SRW stories

34 Stopping time definition Say that T is a stopping time if the event that T = n is in F n for i n.

35 Stopping time definition Say that T is a stopping time if the event that T = n is in F n for i n. In finance applications, T might be the time one sells a stock. Then this states that the decision to sell at time n depends only on prices up to time n, not on (as yet unknown) future prices.

36 Lecture 14 Stopping time examples Let A 1,... be i.i.d. random variables equal to 1 with probability.5 and 1 with probability.5 and let X 0 = 0 and X n = n i=1 A i for n 0.

37 Lecture 14 Stopping time examples Let A 1,... be i.i.d. random variables equal to 1 with probability.5 and 1 with probability.5 and let X 0 = 0 and X n = n i=1 A i for n 0. Which of the following is a stopping time? 1. The smallest T for which X T = The smallest T for which X T { 10, 100} 3. The smallest T for which X T = The T at which the X n sequence achieves the value 17 for the 9th time. 5. The value of T {0, 1, 2,..., 100} for which X T is largest. 6. The largest T {0, 1, 2,..., 100} for which X T = 0.

38 Lecture 14 Stopping time examples Let A 1,... be i.i.d. random variables equal to 1 with probability.5 and 1 with probability.5 and let X 0 = 0 and X n = n i=1 A i for n 0. Which of the following is a stopping time? 1. The smallest T for which X T = The smallest T for which X T { 10, 100} 3. The smallest T for which X T = The T at which the X n sequence achieves the value 17 for the 9th time. 5. The value of T {0, 1, 2,..., 100} for which X T is largest. 6. The largest T {0, 1, 2,..., 100} for which X T = 0.

39 Lecture 14 Stopping time examples Let A 1,... be i.i.d. random variables equal to 1 with probability.5 and 1 with probability.5 and let X 0 = 0 and X n = n i=1 A i for n 0. Which of the following is a stopping time? 1. The smallest T for which X T = The smallest T for which X T { 10, 100} 3. The smallest T for which X T = The T at which the X n sequence achieves the value 17 for the 9th time. 5. The value of T {0, 1, 2,..., 100} for which X T is largest. 6. The largest T {0, 1, 2,..., 100} for which X T = 0. Answer: first four, not last two.

40 Stopping time theorems Theorem: Let X 1, X 2,... be i.i.d. and N a stopping time with N <.

41 Stopping time theorems Theorem: Let X 1, X 2,... be i.i.d. and N a stopping time with N <. Conditioned on stopping time N <, conditional law of {X N+n, n 1} is independent of F n and has same law as original sequence.

42 Stopping time theorems Theorem: Let X 1, X 2,... be i.i.d. and N a stopping time with N <. Conditioned on stopping time N <, conditional law of {X N+n, n 1} is independent of F n and has same law as original sequence. Wald s equation: Let X i be i.i.d. with E X i <. If N is a stopping time with EN < then ES N = EX 1 EN.

43 Stopping time theorems Theorem: Let X 1, X 2,... be i.i.d. and N a stopping time with N <. Conditioned on stopping time N <, conditional law of {X N+n, n 1} is independent of F n and has same law as original sequence. Wald s equation: Let X i be i.i.d. with E X i <. If N is a stopping time with EN < then ES N = EX 1 EN. Wald s second equation: Let X i be i.i.d. with E X i = 0 and EXi 2 = σ 2 <. If N is a stopping time with EN < then ES N = σ 2 EN.

44 Lecture 14 Wald applications to SRW S 0 = a Z and at each time step S j independently changes by ±1 according to a fair coin toss. Fix A Z and let N = inf{k : S k {0, A}. What is ES N?

45 Lecture 14 Wald applications to SRW S 0 = a Z and at each time step S j independently changes by ±1 according to a fair coin toss. Fix A Z and let N = inf{k : S k {0, A}. What is ES N? What is EN?

46 Lecture 14 Outline Infinite divisibility Higher dimensional CFs and CLTs Random walks Stopping times Arcsin law, other SRW stories

47 Lecture 14 Outline Infinite divisibility Higher dimensional CFs and CLTs Random walks Stopping times Arcsin law, other SRW stories

48 Reflection principle How many walks from (0, x) to (n, y) that don t cross the horizontal axis?

49 Reflection principle How many walks from (0, x) to (n, y) that don t cross the horizontal axis? Try counting walks that do cross by giving bijection to walks from (0, x) to (n, y).

50 Lecture 14 Ballot Theorem Suppose that in election candidate A gets α votes and B gets β < α votes. What s probability that A is a head throughout the counting?

51 Lecture 14 Ballot Theorem Suppose that in election candidate A gets α votes and B gets β < α votes. What s probability that A is a head throughout the counting? Answer: (α β)/(α + β). Can be proved using reflection principle.

52 Arcsin theorem Theorem for last hitting time.

53 Arcsin theorem Theorem for last hitting time. Theorem for amount of positive positive time.

Lecture 23. Random walks

Lecture 23. Random walks 18.175: Lecture 23 Random walks Scott Sheffield MIT 1 Outline Random walks Stopping times Arcsin law, other SRW stories 2 Outline Random walks Stopping times Arcsin law, other SRW stories 3 Exchangeable

More information

18.175: Lecture 13 Infinite divisibility and Lévy processes

18.175: Lecture 13 Infinite divisibility and Lévy processes 18.175 Lecture 13 18.175: Lecture 13 Infinite divisibility and Lévy processes Scott Sheffield MIT Outline Poisson random variable convergence Extend CLT idea to stable random variables Infinite divisibility

More information

18.175: Lecture 17 Poisson random variables

18.175: Lecture 17 Poisson random variables 18.175: Lecture 17 Poisson random variables Scott Sheffield MIT 1 Outline More on random walks and local CLT Poisson random variable convergence Extend CLT idea to stable random variables 2 Outline More

More information

18.175: Lecture 2 Extension theorems, random variables, distributions

18.175: Lecture 2 Extension theorems, random variables, distributions 18.175: Lecture 2 Extension theorems, random variables, distributions Scott Sheffield MIT Outline Extension theorems Characterizing measures on R d Random variables Outline Extension theorems Characterizing

More information

18.440: Lecture 19 Normal random variables

18.440: Lecture 19 Normal random variables 18.440 Lecture 19 18.440: Lecture 19 Normal random variables Scott Sheffield MIT Outline Tossing coins Normal random variables Special case of central limit theorem Outline Tossing coins Normal random

More information

Lecture 2. We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales.

Lecture 2. We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales. Lecture 2 1 Martingales We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales. 1.1 Doob s inequality We have the following maximal

More information

Martingale Theory for Finance

Martingale Theory for Finance Martingale Theory for Finance Tusheng Zhang October 27, 2015 1 Introduction 2 Probability spaces and σ-fields 3 Integration with respect to a probability measure. 4 Conditional expectation. 5 Martingales.

More information

18.175: Lecture 8 Weak laws and moment-generating/characteristic functions

18.175: Lecture 8 Weak laws and moment-generating/characteristic functions 18.175: Lecture 8 Weak laws and moment-generating/characteristic functions Scott Sheffield MIT 18.175 Lecture 8 1 Outline Moment generating functions Weak law of large numbers: Markov/Chebyshev approach

More information

Notes 15 : UI Martingales

Notes 15 : UI Martingales Notes 15 : UI Martingales Math 733 - Fall 2013 Lecturer: Sebastien Roch References: [Wil91, Chapter 13, 14], [Dur10, Section 5.5, 5.6, 5.7]. 1 Uniform Integrability We give a characterization of L 1 convergence.

More information

1 Stat 605. Homework I. Due Feb. 1, 2011

1 Stat 605. Homework I. Due Feb. 1, 2011 The first part is homework which you need to turn in. The second part is exercises that will not be graded, but you need to turn it in together with the take-home final exam. 1 Stat 605. Homework I. Due

More information

Lecture Notes 3 Convergence (Chapter 5)

Lecture Notes 3 Convergence (Chapter 5) Lecture Notes 3 Convergence (Chapter 5) 1 Convergence of Random Variables Let X 1, X 2,... be a sequence of random variables and let X be another random variable. Let F n denote the cdf of X n and let

More information

18.440: Lecture 28 Lectures Review

18.440: Lecture 28 Lectures Review 18.440: Lecture 28 Lectures 18-27 Review Scott Sheffield MIT Outline Outline It s the coins, stupid Much of what we have done in this course can be motivated by the i.i.d. sequence X i where each X i is

More information

Lecture 7. Sums of random variables

Lecture 7. Sums of random variables 18.175: Lecture 7 Sums of random variables Scott Sheffield MIT 18.175 Lecture 7 1 Outline Definitions Sums of random variables 18.175 Lecture 7 2 Outline Definitions Sums of random variables 18.175 Lecture

More information

18.440: Lecture 28 Lectures Review

18.440: Lecture 28 Lectures Review 18.440: Lecture 28 Lectures 17-27 Review Scott Sheffield MIT 1 Outline Continuous random variables Problems motivated by coin tossing Random variable properties 2 Outline Continuous random variables Problems

More information

Probability Theory I: Syllabus and Exercise

Probability Theory I: Syllabus and Exercise Probability Theory I: Syllabus and Exercise Narn-Rueih Shieh **Copyright Reserved** This course is suitable for those who have taken Basic Probability; some knowledge of Real Analysis is recommended( will

More information

Problem Sheet 1. You may assume that both F and F are σ-fields. (a) Show that F F is not a σ-field. (b) Let X : Ω R be defined by 1 if n = 1

Problem Sheet 1. You may assume that both F and F are σ-fields. (a) Show that F F is not a σ-field. (b) Let X : Ω R be defined by 1 if n = 1 Problem Sheet 1 1. Let Ω = {1, 2, 3}. Let F = {, {1}, {2, 3}, {1, 2, 3}}, F = {, {2}, {1, 3}, {1, 2, 3}}. You may assume that both F and F are σ-fields. (a) Show that F F is not a σ-field. (b) Let X :

More information

JUSTIN HARTMANN. F n Σ.

JUSTIN HARTMANN. F n Σ. BROWNIAN MOTION JUSTIN HARTMANN Abstract. This paper begins to explore a rigorous introduction to probability theory using ideas from algebra, measure theory, and other areas. We start with a basic explanation

More information

MAT 135B Midterm 1 Solutions

MAT 135B Midterm 1 Solutions MAT 35B Midterm Solutions Last Name (PRINT): First Name (PRINT): Student ID #: Section: Instructions:. Do not open your test until you are told to begin. 2. Use a pen to print your name in the spaces above.

More information

Math489/889 Stochastic Processes and Advanced Mathematical Finance Solutions for Homework 7

Math489/889 Stochastic Processes and Advanced Mathematical Finance Solutions for Homework 7 Math489/889 Stochastic Processes and Advanced Mathematical Finance Solutions for Homework 7 Steve Dunbar Due Mon, November 2, 2009. Time to review all of the information we have about coin-tossing fortunes

More information

Lecture 12. Poisson random variables

Lecture 12. Poisson random variables 18.440: Lecture 12 Poisson random variables Scott Sheffield MIT 1 Outline Poisson random variable definition Poisson random variable properties Poisson random variable problems 2 Outline Poisson random

More information

Lecture 14. More discrete random variables

Lecture 14. More discrete random variables 18.440: Lecture 14 More discrete random variables Scott Sheffield MIT 1 Outline Geometric random variables Negative binomial random variables Problems 2 Outline Geometric random variables Negative binomial

More information

IEOR 3106: Introduction to Operations Research: Stochastic Models. Professor Whitt. SOLUTIONS to Homework Assignment 2

IEOR 3106: Introduction to Operations Research: Stochastic Models. Professor Whitt. SOLUTIONS to Homework Assignment 2 IEOR 316: Introduction to Operations Research: Stochastic Models Professor Whitt SOLUTIONS to Homework Assignment 2 More Probability Review: In the Ross textbook, Introduction to Probability Models, read

More information

Lecture 9. Expectations of discrete random variables

Lecture 9. Expectations of discrete random variables 18.440: Lecture 9 Expectations of discrete random variables Scott Sheffield MIT 1 Outline Defining expectation Functions of random variables Motivation 2 Outline Defining expectation Functions of random

More information

18.175: Lecture 3 Integration

18.175: Lecture 3 Integration 18.175: Lecture 3 Scott Sheffield MIT Outline Outline Recall definitions Probability space is triple (Ω, F, P) where Ω is sample space, F is set of events (the σ-algebra) and P : F [0, 1] is the probability

More information

On Optimal Stopping Problems with Power Function of Lévy Processes

On Optimal Stopping Problems with Power Function of Lévy Processes On Optimal Stopping Problems with Power Function of Lévy Processes Budhi Arta Surya Department of Mathematics University of Utrecht 31 August 2006 This talk is based on the joint paper with A.E. Kyprianou:

More information

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015 Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.

More information

Lecture 17 Brownian motion as a Markov process

Lecture 17 Brownian motion as a Markov process Lecture 17: Brownian motion as a Markov process 1 of 14 Course: Theory of Probability II Term: Spring 2015 Instructor: Gordan Zitkovic Lecture 17 Brownian motion as a Markov process Brownian motion is

More information

Example continued. Math 425 Intro to Probability Lecture 37. Example continued. Example

Example continued. Math 425 Intro to Probability Lecture 37. Example continued. Example continued : Coin tossing Math 425 Intro to Probability Lecture 37 Kenneth Harris kaharri@umich.edu Department of Mathematics University of Michigan April 8, 2009 Consider a Bernoulli trials process with

More information

18.600: Lecture 7 Bayes formula and independence

18.600: Lecture 7 Bayes formula and independence 18.600: Lecture 7 Bayes formula and independence Scott Sheffield MIT Outline Bayes formula Independence Outline Bayes formula Independence Recall definition: conditional probability Definition: P(E F )

More information

STA205 Probability: Week 8 R. Wolpert

STA205 Probability: Week 8 R. Wolpert INFINITE COIN-TOSS AND THE LAWS OF LARGE NUMBERS The traditional interpretation of the probability of an event E is its asymptotic frequency: the limit as n of the fraction of n repeated, similar, and

More information

Problem Points S C O R E Total: 120

Problem Points S C O R E Total: 120 PSTAT 160 A Final Exam December 10, 2015 Name Student ID # Problem Points S C O R E 1 10 2 10 3 10 4 10 5 10 6 10 7 10 8 10 9 10 10 10 11 10 12 10 Total: 120 1. (10 points) Take a Markov chain with the

More information

18.175: Lecture 15 Characteristic functions and central limit theorem

18.175: Lecture 15 Characteristic functions and central limit theorem 18.175: Lecture 15 Characteristic functions and central limit theorem Scott Sheffield MIT Outline Characteristic functions Outline Characteristic functions Characteristic functions Let X be a random variable.

More information

Lecture 9: Conditional Probability and Independence

Lecture 9: Conditional Probability and Independence EE5110: Probability Foundations July-November 2015 Lecture 9: Conditional Probability and Independence Lecturer: Dr. Krishna Jagannathan Scribe: Vishakh Hegde 9.1 Conditional Probability Definition 9.1

More information

Lecture 7. Bayes formula and independence

Lecture 7. Bayes formula and independence 18.440: Lecture 7 Bayes formula and independence Scott Sheffield MIT 1 Outline Bayes formula Independence 2 Outline Bayes formula Independence 3 Recall definition: conditional probability Definition: P(E

More information

On the definition and properties of p-harmonious functions

On the definition and properties of p-harmonious functions On the definition and properties of p-harmonious functions University of Pittsburgh, UBA, UAM Workshop on New Connections Between Differential and Random Turn Games, PDE s and Image Processing Pacific

More information

Lecture 9. d N(0, 1). Now we fix n and think of a SRW on [0,1]. We take the k th step at time k n. and our increments are ± 1

Lecture 9. d N(0, 1). Now we fix n and think of a SRW on [0,1]. We take the k th step at time k n. and our increments are ± 1 Random Walks and Brownian Motion Tel Aviv University Spring 011 Lecture date: May 0, 011 Lecture 9 Instructor: Ron Peled Scribe: Jonathan Hermon In today s lecture we present the Brownian motion (BM).

More information

Mod-φ convergence I: examples and probabilistic estimates

Mod-φ convergence I: examples and probabilistic estimates Mod-φ convergence I: examples and probabilistic estimates Valentin Féray (joint work with Pierre-Loïc Méliot and Ashkan Nikeghbali) Institut für Mathematik, Universität Zürich Summer school in Villa Volpi,

More information

18.600: Lecture 7 Bayes formula and independence

18.600: Lecture 7 Bayes formula and independence 18.600 Lecture 7 18.600: Lecture 7 Bayes formula and independence Scott Sheffield MIT 18.600 Lecture 7 Outline Bayes formula Independence 18.600 Lecture 7 Outline Bayes formula Independence Recall definition:

More information

Problem Points S C O R E Total: 120

Problem Points S C O R E Total: 120 PSTAT 160 A Final Exam Solution December 10, 2015 Name Student ID # Problem Points S C O R E 1 10 2 10 3 10 4 10 5 10 6 10 7 10 8 10 9 10 10 10 11 10 12 10 Total: 120 1. (10 points) Take a Markov chain

More information

Elementary Probability. Exam Number 38119

Elementary Probability. Exam Number 38119 Elementary Probability Exam Number 38119 2 1. Introduction Consider any experiment whose result is unknown, for example throwing a coin, the daily number of customers in a supermarket or the duration of

More information

18.440: Lecture 39 Review: practice problems

18.440: Lecture 39 Review: practice problems 18.440 Lecture 39 18.440: Lecture 39 Review: practice problems Scott Sheffield MIT 18.440 Lecture 39 Markov chains Alice and Bob share a home with a bathroom, a walk-in closet, and 2 towels. 18.440 Lecture

More information

1 Presessional Probability

1 Presessional Probability 1 Presessional Probability Probability theory is essential for the development of mathematical models in finance, because of the randomness nature of price fluctuations in the markets. This presessional

More information

Stat 100a: Introduction to Probability. Outline for the day:

Stat 100a: Introduction to Probability. Outline for the day: Stat 100a: Introduction to Probability. Outline for the day: 1. Midterm2 back. 2. CLT again. 3. Confidence intervals. 4. Sample size calculation. 5. Random walks. 6. Reflection principle. 7. Ballot theorem.

More information

X = X X n, + X 2

X = X X n, + X 2 CS 70 Discrete Mathematics for CS Fall 2003 Wagner Lecture 22 Variance Question: At each time step, I flip a fair coin. If it comes up Heads, I walk one step to the right; if it comes up Tails, I walk

More information

Filtrations, Markov Processes and Martingales. Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition

Filtrations, Markov Processes and Martingales. Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition Filtrations, Markov Processes and Martingales Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition David pplebaum Probability and Statistics Department,

More information

Statistical Inference

Statistical Inference Statistical Inference Lecture 1: Probability Theory MING GAO DASE @ ECNU (for course related communications) mgao@dase.ecnu.edu.cn Sep. 11, 2018 Outline Introduction Set Theory Basics of Probability Theory

More information

The Central Limit Theorem: More of the Story

The Central Limit Theorem: More of the Story The Central Limit Theorem: More of the Story Steven Janke November 2015 Steven Janke (Seminar) The Central Limit Theorem:More of the Story November 2015 1 / 33 Central Limit Theorem Theorem (Central Limit

More information

18440: Probability and Random variables Quiz 1 Friday, October 17th, 2014

18440: Probability and Random variables Quiz 1 Friday, October 17th, 2014 18440: Probability and Random variables Quiz 1 Friday, October 17th, 014 You will have 55 minutes to complete this test. Each of the problems is worth 5 % of your exam grade. No calculators, notes, or

More information

Martingale Theory and Applications

Martingale Theory and Applications Martingale Theory and Applications Dr Nic Freeman June 4, 2015 Contents 1 Conditional Expectation 2 1.1 Probability spaces and σ-fields............................ 2 1.2 Random Variables...................................

More information

Expected Value 7/7/2006

Expected Value 7/7/2006 Expected Value 7/7/2006 Definition Let X be a numerically-valued discrete random variable with sample space Ω and distribution function m(x). The expected value E(X) is defined by E(X) = x Ω x m(x), provided

More information

RANDOM WALKS AND THE PROBABILITY OF RETURNING HOME

RANDOM WALKS AND THE PROBABILITY OF RETURNING HOME RANDOM WALKS AND THE PROBABILITY OF RETURNING HOME ELIZABETH G. OMBRELLARO Abstract. This paper is expository in nature. It intuitively explains, using a geometrical and measure theory perspective, why

More information

Advanced Probability Theory (Math541)

Advanced Probability Theory (Math541) Advanced Probability Theory (Math541) Instructor: Kani Chen (Classic)/Modern Probability Theory (1900-1960) Instructor: Kani Chen (HKUST) Advanced Probability Theory (Math541) 1 / 17 Primitive/Classic

More information

IEOR 6711: Stochastic Models I SOLUTIONS to the First Midterm Exam, October 7, 2008

IEOR 6711: Stochastic Models I SOLUTIONS to the First Midterm Exam, October 7, 2008 IEOR 6711: Stochastic Models I SOLUTIONS to the First Midterm Exam, October 7, 2008 Justify your answers; show your work. 1. A sequence of Events. (10 points) Let {B n : n 1} be a sequence of events in

More information

18.600: Lecture 39 Review: practice problems

18.600: Lecture 39 Review: practice problems 18.600: Lecture 39 Review: practice problems Scott Sheffield MIT Markov chains Alice and Bob share a home with a bathroom, a walk-in closet, and 2 towels. Markov chains Alice and Bob share a home with

More information

8 Laws of large numbers

8 Laws of large numbers 8 Laws of large numbers 8.1 Introduction We first start with the idea of standardizing a random variable. Let X be a random variable with mean µ and variance σ 2. Then Z = (X µ)/σ will be a random variable

More information

Lecture 16. Lectures 1-15 Review

Lecture 16. Lectures 1-15 Review 18.440: Lecture 16 Lectures 1-15 Review Scott Sheffield MIT 1 Outline Counting tricks and basic principles of probability Discrete random variables 2 Outline Counting tricks and basic principles of probability

More information

UC Berkeley Department of Electrical Engineering and Computer Sciences. EECS 126: Probability and Random Processes

UC Berkeley Department of Electrical Engineering and Computer Sciences. EECS 126: Probability and Random Processes UC Berkeley Department of Electrical Engineering and Computer Sciences EECS 26: Probability and Random Processes Problem Set Spring 209 Self-Graded Scores Due:.59 PM, Monday, February 4, 209 Submit your

More information

A PECULIAR COIN-TOSSING MODEL

A PECULIAR COIN-TOSSING MODEL A PECULIAR COIN-TOSSING MODEL EDWARD J. GREEN 1. Coin tossing according to de Finetti A coin is drawn at random from a finite set of coins. Each coin generates an i.i.d. sequence of outcomes (heads or

More information

Notes 9 : Infinitely divisible and stable laws

Notes 9 : Infinitely divisible and stable laws Notes 9 : Infinitely divisible and stable laws Math 733 - Fall 203 Lecturer: Sebastien Roch References: [Dur0, Section 3.7, 3.8], [Shi96, Section III.6]. Infinitely divisible distributions Recall: EX 9.

More information

Lectures on Stochastic Stability. Sergey FOSS. Heriot-Watt University. Lecture 4. Coupling and Harris Processes

Lectures on Stochastic Stability. Sergey FOSS. Heriot-Watt University. Lecture 4. Coupling and Harris Processes Lectures on Stochastic Stability Sergey FOSS Heriot-Watt University Lecture 4 Coupling and Harris Processes 1 A simple example Consider a Markov chain X n in a countable state space S with transition probabilities

More information

Probability and Measure

Probability and Measure Part II Year 2018 2017 2016 2015 2014 2013 2012 2011 2010 2009 2008 2007 2006 2005 2018 84 Paper 4, Section II 26J Let (X, A) be a measurable space. Let T : X X be a measurable map, and µ a probability

More information

Lecture 6: The Pigeonhole Principle and Probability Spaces

Lecture 6: The Pigeonhole Principle and Probability Spaces Lecture 6: The Pigeonhole Principle and Probability Spaces Anup Rao January 17, 2018 We discuss the pigeonhole principle and probability spaces. Pigeonhole Principle The pigeonhole principle is an extremely

More information

1 Sequences of events and their limits

1 Sequences of events and their limits O.H. Probability II (MATH 2647 M15 1 Sequences of events and their limits 1.1 Monotone sequences of events Sequences of events arise naturally when a probabilistic experiment is repeated many times. For

More information

A D VA N C E D P R O B A B I L - I T Y

A D VA N C E D P R O B A B I L - I T Y A N D R E W T U L L O C H A D VA N C E D P R O B A B I L - I T Y T R I N I T Y C O L L E G E T H E U N I V E R S I T Y O F C A M B R I D G E Contents 1 Conditional Expectation 5 1.1 Discrete Case 6 1.2

More information

18.600: Lecture 24 Covariance and some conditional expectation exercises

18.600: Lecture 24 Covariance and some conditional expectation exercises 18.600: Lecture 24 Covariance and some conditional expectation exercises Scott Sheffield MIT Outline Covariance and correlation Paradoxes: getting ready to think about conditional expectation Outline Covariance

More information

1 Poisson processes, and Compound (batch) Poisson processes

1 Poisson processes, and Compound (batch) Poisson processes Copyright c 2007 by Karl Sigman 1 Poisson processes, and Compound (batch) Poisson processes 1.1 Point Processes Definition 1.1 A simple point process ψ = {t n : n 1} is a sequence of strictly increasing

More information

Lecture 1: August 28

Lecture 1: August 28 36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 1: August 28 Our broad goal for the first few lectures is to try to understand the behaviour of sums of independent random

More information

Exercises with solutions (Set D)

Exercises with solutions (Set D) Exercises with solutions Set D. A fair die is rolled at the same time as a fair coin is tossed. Let A be the number on the upper surface of the die and let B describe the outcome of the coin toss, where

More information

1 Basic continuous random variable problems

1 Basic continuous random variable problems Name M362K Final Here are problems concerning material from Chapters 5 and 6. To review the other chapters, look over previous practice sheets for the two exams, previous quizzes, previous homeworks and

More information

Random Models. Tusheng Zhang. February 14, 2013

Random Models. Tusheng Zhang. February 14, 2013 Random Models Tusheng Zhang February 14, 013 1 Introduction In this module, we will introduce some random models which have many real life applications. The course consists of four parts. 1. A brief review

More information

Chapter 1. Sets and probability. 1.3 Probability space

Chapter 1. Sets and probability. 1.3 Probability space Random processes - Chapter 1. Sets and probability 1 Random processes Chapter 1. Sets and probability 1.3 Probability space 1.3 Probability space Random processes - Chapter 1. Sets and probability 2 Probability

More information

Infinitely divisible distributions and the Lévy-Khintchine formula

Infinitely divisible distributions and the Lévy-Khintchine formula Infinitely divisible distributions and the Cornell University May 1, 2015 Some definitions Let X be a real-valued random variable with law µ X. Recall that X is said to be infinitely divisible if for every

More information

A = A U. U [n] P(A U ). n 1. 2 k(n k). k. k=1

A = A U. U [n] P(A U ). n 1. 2 k(n k). k. k=1 Lecture I jacques@ucsd.edu Notation: Throughout, P denotes probability and E denotes expectation. Denote (X) (r) = X(X 1)... (X r + 1) and let G n,p denote the Erdős-Rényi model of random graphs. 10 Random

More information

Lecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN

Lecture Notes 5 Convergence and Limit Theorems. Convergence with Probability 1. Convergence in Mean Square. Convergence in Probability, WLLN Lecture Notes 5 Convergence and Limit Theorems Motivation Convergence with Probability Convergence in Mean Square Convergence in Probability, WLLN Convergence in Distribution, CLT EE 278: Convergence and

More information

Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016

Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016 8. For any two events E and F, P (E) = P (E F ) + P (E F c ). Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016 Sample space. A sample space consists of a underlying

More information

Probability Theory. Richard F. Bass

Probability Theory. Richard F. Bass Probability Theory Richard F. Bass ii c Copyright 2014 Richard F. Bass Contents 1 Basic notions 1 1.1 A few definitions from measure theory............. 1 1.2 Definitions............................. 2

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 3 9/10/2008 CONDITIONING AND INDEPENDENCE

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 3 9/10/2008 CONDITIONING AND INDEPENDENCE MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 3 9/10/2008 CONDITIONING AND INDEPENDENCE Most of the material in this lecture is covered in [Bertsekas & Tsitsiklis] Sections 1.3-1.5

More information

Chapter 1 Probability Theory

Chapter 1 Probability Theory Review for the previous lecture Eample: how to calculate probabilities of events (especially for sampling with replacement) and the conditional probability Definition: conditional probability, statistically

More information

Probability and Measure

Probability and Measure Chapter 4 Probability and Measure 4.1 Introduction In this chapter we will examine probability theory from the measure theoretic perspective. The realisation that measure theory is the foundation of probability

More information

Random Process Lecture 1. Fundamentals of Probability

Random Process Lecture 1. Fundamentals of Probability Random Process Lecture 1. Fundamentals of Probability Husheng Li Min Kao Department of Electrical Engineering and Computer Science University of Tennessee, Knoxville Spring, 2016 1/43 Outline 2/43 1 Syllabus

More information

Part II Probability and Measure

Part II Probability and Measure Part II Probability and Measure Theorems Based on lectures by J. Miller Notes taken by Dexter Chua Michaelmas 2016 These notes are not endorsed by the lecturers, and I have modified them (often significantly)

More information

CMPSCI 240: Reasoning Under Uncertainty

CMPSCI 240: Reasoning Under Uncertainty CMPSCI 240: Reasoning Under Uncertainty Lecture 5 Prof. Hanna Wallach wallach@cs.umass.edu February 7, 2012 Reminders Pick up a copy of B&T Check the course website: http://www.cs.umass.edu/ ~wallach/courses/s12/cmpsci240/

More information

5 Birkhoff s Ergodic Theorem

5 Birkhoff s Ergodic Theorem 5 Birkhoff s Ergodic Theorem Birkhoff s Ergodic Theorem extends the validity of Kolmogorov s strong law to the class of stationary sequences of random variables. Stationary sequences occur naturally even

More information

Lecture 10. Variance and standard deviation

Lecture 10. Variance and standard deviation 18.440: Lecture 10 Variance and standard deviation Scott Sheffield MIT 1 Outline Defining variance Examples Properties Decomposition trick 2 Outline Defining variance Examples Properties Decomposition

More information

1* (10 pts) Let X be a random variable with P (X = 1) = P (X = 1) = 1 2

1* (10 pts) Let X be a random variable with P (X = 1) = P (X = 1) = 1 2 Math 736-1 Homework Fall 27 1* (1 pts) Let X be a random variable with P (X = 1) = P (X = 1) = 1 2 and let Y be a standard normal random variable. Assume that X and Y are independent. Find the distribution

More information

Lecture 8: Continuous random variables, expectation and variance

Lecture 8: Continuous random variables, expectation and variance Lecture 8: Continuous random variables, expectation and variance Lejla Batina Institute for Computing and Information Sciences Digital Security Version: autumn 2013 Lejla Batina Version: autumn 2013 Wiskunde

More information

Randomization in Algorithms

Randomization in Algorithms Randomization in Algorithms Randomization is a tool for designing good algorithms. Two kinds of algorithms Las Vegas - always correct, running time is random. Monte Carlo - may return incorrect answers,

More information

Math 564 Homework 1. Solutions.

Math 564 Homework 1. Solutions. Math 564 Homework 1. Solutions. Problem 1. Prove Proposition 0.2.2. A guide to this problem: start with the open set S = (a, b), for example. First assume that a >, and show that the number a has the properties

More information

6.041/6.431 Fall 2010 Quiz 2 Solutions

6.041/6.431 Fall 2010 Quiz 2 Solutions 6.04/6.43: Probabilistic Systems Analysis (Fall 200) 6.04/6.43 Fall 200 Quiz 2 Solutions Problem. (80 points) In this problem: (i) X is a (continuous) uniform random variable on [0, 4]. (ii) Y is an exponential

More information

Lecture 8 Sampling Theory

Lecture 8 Sampling Theory Lecture 8 Sampling Theory Thais Paiva STA 111 - Summer 2013 Term II July 11, 2013 1 / 25 Thais Paiva STA 111 - Summer 2013 Term II Lecture 8, 07/11/2013 Lecture Plan 1 Sampling Distributions 2 Law of Large

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 7 9/25/2013

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 7 9/25/2013 MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.65/15.070J Fall 013 Lecture 7 9/5/013 The Reflection Principle. The Distribution of the Maximum. Brownian motion with drift Content. 1. Quick intro to stopping times.

More information

I. ANALYSIS; PROBABILITY

I. ANALYSIS; PROBABILITY ma414l1.tex Lecture 1. 12.1.2012 I. NLYSIS; PROBBILITY 1. Lebesgue Measure and Integral We recall Lebesgue measure (M411 Probability and Measure) λ: defined on intervals (a, b] by λ((a, b]) := b a (so

More information

1: PROBABILITY REVIEW

1: PROBABILITY REVIEW 1: PROBABILITY REVIEW Marek Rutkowski School of Mathematics and Statistics University of Sydney Semester 2, 2016 M. Rutkowski (USydney) Slides 1: Probability Review 1 / 56 Outline We will review the following

More information

MATH MW Elementary Probability Course Notes Part I: Models and Counting

MATH MW Elementary Probability Course Notes Part I: Models and Counting MATH 2030 3.00MW Elementary Probability Course Notes Part I: Models and Counting Tom Salisbury salt@yorku.ca York University Winter 2010 Introduction [Jan 5] Probability: the mathematics used for Statistics

More information

1 Probability theory. 2 Random variables and probability theory.

1 Probability theory. 2 Random variables and probability theory. Probability theory Here we summarize some of the probability theory we need. If this is totally unfamiliar to you, you should look at one of the sources given in the readings. In essence, for the major

More information

Lecture 1. ABC of Probability

Lecture 1. ABC of Probability Math 408 - Mathematical Statistics Lecture 1. ABC of Probability January 16, 2013 Konstantin Zuev (USC) Math 408, Lecture 1 January 16, 2013 1 / 9 Agenda Sample Spaces Realizations, Events Axioms of Probability

More information

Lecture 5: Asymptotic Equipartition Property

Lecture 5: Asymptotic Equipartition Property Lecture 5: Asymptotic Equipartition Property Law of large number for product of random variables AEP and consequences Dr. Yao Xie, ECE587, Information Theory, Duke University Stock market Initial investment

More information

6.262: Discrete Stochastic Processes 2/2/11. Lecture 1: Introduction and Probability review

6.262: Discrete Stochastic Processes 2/2/11. Lecture 1: Introduction and Probability review 6.262: Discrete Stochastic Processes 2/2/11 Lecture 1: Introduction and Probability review Outline: Probability in the real world Probability as a branch of mathematics Discrete stochastic processes Processes

More information

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015 Part IA Probability Theorems Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.

More information

1 Variance of a Random Variable

1 Variance of a Random Variable Indian Institute of Technology Bombay Department of Electrical Engineering Handout 14 EE 325 Probability and Random Processes Lecture Notes 9 August 28, 2014 1 Variance of a Random Variable The expectation

More information