Math-Stat-491-Fall2014-Notes-V

Similar documents
Math-Stat-491-Fall2014-Notes-I

Inference for Stochastic Processes

Markov Chains (Part 3)

The Boundary Problem: Markov Chain Solution

Martingale Theory for Finance

Probability Models. 4. What is the definition of the expectation of a discrete random variable?

Lecture 4 - Random walk, ruin problems and random processes

Lecture 4. David Aldous. 2 September David Aldous Lecture 4

Lectures on Markov Chains

T 1. The value function v(x) is the expected net gain when using the optimal stopping time starting at state x:

Week 12-13: Discrete Probability

RANDOM WALKS IN ONE DIMENSION

General Boundary Problem: System of Equations Solution

CS 361: Probability & Statistics

Notes 13 : Conditioning

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

EECS 126 Probability and Random Processes University of California, Berkeley: Spring 2015 Abhay Parekh February 17, 2015.

MATH HOMEWORK PROBLEMS D. MCCLENDON

Relations and Equivalence Relations

Math 510 midterm 3 answers

Math 456: Mathematical Modeling. Tuesday, March 6th, 2018

Lecture 22: Variance and Covariance

1 Variance of a Random Variable

Lecture 9 Classification of States

Course: ESO-209 Home Work: 1 Instructor: Debasis Kundu

MARTINGALE APPROXIMATIONS FOR ERGODIC SYSTEMS

RANDOM WALKS IN Z d AND THE DIRICHLET PROBLEM

Lecture 12: Multiple Random Variables and Independence

1 Random Walks and Electrical Networks

1 Gambler s Ruin Problem

More than one variable

A D VA N C E D P R O B A B I L - I T Y

4 Branching Processes

Math 6810 (Probability) Fall Lecture notes

Name: 180A MIDTERM 2. (x + n)/2

Measures and Measure Spaces

. Find E(V ) and var(v ).

Probability. Computer Science Tripos, Part IA. R.J. Gibbens. Computer Laboratory University of Cambridge. Easter Term 2008/9

MATH 56A SPRING 2008 STOCHASTIC PROCESSES

Lecture 9. d N(0, 1). Now we fix n and think of a SRW on [0,1]. We take the k th step at time k n. and our increments are ± 1

Random walk on a polygon

Probability, Random Processes and Inference

Stochastic process for macro

Stat-491-Fall2014-Assignment-III

CONVERGENCE OF RANDOM SERIES AND MARTINGALES

Connectedness. Proposition 2.2. The following are equivalent for a topological space (X, T ).

Math 564 Homework 1. Solutions.

Continuum Probability and Sets of Measure Zero

Discrete Probability Refresher

Lecture Notes 7 Random Processes. Markov Processes Markov Chains. Random Processes

MAS113 Introduction to Probability and Statistics. Proofs of theorems

ADVANCED PROBABILITY: SOLUTIONS TO SHEET 1

SOLVING AND COMPUTING THE DISCRETE DIRICHLET PROBLEM

ECE 302, Final 3:20-5:20pm Mon. May 1, WTHR 160 or WTHR 172.

Some Definition and Example of Markov Chain

Modern Discrete Probability Branching processes

Introduction to Stochastic Processes

JOSEPH B. KADANE AND JIASHUN JIN, IN HONOR OF ARNOLD ZELLNER

3. DISCRETE RANDOM VARIABLES

P (A G) dp G P (A G)

Lectures 22-23: Conditional Expectations

1 Presessional Probability

Stat 516, Homework 1

Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016

Discrete Mathematics and Probability Theory Fall 2014 Anant Sahai Note 15. Random Variables: Distributions, Independence, and Expectations

Synthetic Probability Theory

Math Homework 5 Solutions

4 Sums of Independent Random Variables

Expectation is linear. So far we saw that E(X + Y ) = E(X) + E(Y ). Let α R. Then,

Report 1 The Axiom of Choice

Simultaneous Equations Solve for x and y (What are the values of x and y): Summation What is the value of the following given x = j + 1. x i.

Lecture 1: Brief Review on Stochastic Processes

Mathematical Methods for Computer Science

Markov Chains CK eqns Classes Hitting times Rec./trans. Strong Markov Stat. distr. Reversibility * Markov Chains

Arkansas Tech University MATH 3513: Applied Statistics I Dr. Marcel B. Finan

Math 381 Discrete Mathematical Modeling

Lecture 4: Probability and Discrete Random Variables

Expected Value 7/7/2006

Lecture 5. If we interpret the index n 0 as time, then a Markov chain simply requires that the future depends only on the present and not on the past.

HW 2 Solutions. The variance of the random walk is explosive (lim n Var (X n ) = ).

STAT2201. Analysis of Engineering & Scientific Data. Unit 3

1 Stat 605. Homework I. Due Feb. 1, 2011

Business Statistics 41000: Homework # 2 Solutions

Monte Carlo Integration II & Sampling from PDFs

Lecture 4 An Introduction to Stochastic Processes

Problem Sheet 1. You may assume that both F and F are σ-fields. (a) Show that F F is not a σ-field. (b) Let X : Ω R be defined by 1 if n = 1

RANDOM WALKS AND THE PROBABILITY OF RETURNING HOME

Exercise Exercise Homework #6 Solutions Thursday 6 April 2006

Euclidean Space. This is a brief review of some basic concepts that I hope will already be familiar to you.

We will briefly look at the definition of a probability space, probability measures, conditional probability and independence of probability events.

MATH 564/STAT 555 Applied Stochastic Processes Homework 2, September 18, 2015 Due September 30, 2015

EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY

THE QUEEN S UNIVERSITY OF BELFAST

Lesson 11: The Special Role of Zero in Factoring

Polytechnic Institute of NYU MA 2212 MIDTERM Feb 12, 2009

Probability Theory and Statistics. Peter Jochumzen

MATH 56A: STOCHASTIC PROCESSES CHAPTER 2

Random Models. Tusheng Zhang. February 14, 2013

Course Notes. Part IV. Probabilistic Combinatorics. Algorithms

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Transcription:

Math-Stat-491-Fall2014-Notes-V Hariharan Narayanan November 18, 2014 Martingales 1 Introduction Martingales were originally introduced into probability theory as a model for fair betting games. Essentially we bet on events of known probability according to these known values and the payoff is also according to these values. Common sense tells us that in that case, in the long run, we should neither win nor lose. The theory more or less captures this idea but there are paradoxes - for instance if you play until you are ahead you will gain but you might need unbounded resources to reach that stage. However, the value of martingale theory far exceeds the original reason for their coming into being. These notes are intended to atleast partially justify this statement. 2 Preliminaries The theory of martingales makes repeated use of the notion of conditional expectation. We review the theory briefly. The conditional expectation of a random variable X given that a random variable Y = y, is given by E[X {Y = y}] x p X Y (x y)dx. In the discrete case this becomes E[X {Y = y}] Σ i x i p X Y (x i y). Therefore we have E[X] = Σ i x i p X (x i ) = Σ j [Σ i x i p X Y (x i y j )] p Y (y j ). Now the expression E[X {Y = y}] defines a function of y. Consider the function f( ) on the sample space defined by f(ω) E[X {Y = Y (ω)}]. To compute the value of f( ) on a point ω in the sample space, we first compute Y (ω), then compute the expectation E[X {Y = Y (ω)}]. This random variable f( ) is usually denoted E[X Y ]. Similarly, the random variable (E[X Y ]) 2 is simply the square of the random variable E[X Y ]. 1

Note that, while X {Y = y} is a random variable whose sample space is the set of all ω which map to y under the random variable Y, the expression X Y does not have a predetermined meaning. Next, what meaning should we assign V ar[x Y ]? Formally, V ar[z] E[Z 2 ] (E[Z]) 2. So we define V ar[x Y ] E[X 2 Y ] (E[X Y ]) 2. Thus V ar[x Y ] is the random variable given by the expression on the right hand side. We will now prove the law of total expectation: E[E[X Y ]] of the random variable E[X Y ] is equal to E[X]. In the discrete case, we could first partition the sample space into preimages under Y of the values Y (ω) = y i, find the probability associated with each set, multiply it by E[X {Y = y i }] and then compute the sum of all such products. Thus we get E[E[X Y ]] = Σ i E[X {Y = y i }] p Y (y i ). Hence E[E[X Y ]] = Σ i Σ j x j p X Y (x j y i ) p Y (y i ). This, by using definition of conditional probability becomes E[E[X Y ]] = Σ i Σ j x j p XY (x j, y i ). Interchanging summation we get E[E[X Y ]] = Σ j Σ i x j p XY (x j, y i ). We can rewrite this as i.e., as E[E[X Y ]] = Σ j x j [Σ i p XY (x j, y i )], E[E[X Y ]] = Σ j x j p X (x j )] = E[X]. In martingale theory we also come across expressions of the kind E[E[X Y, Z] Z] and a useful result is a form of the law of total expectation: E[X Z] = E[E[X Y, Z] Z]. We will call this the second form of the law of total expectation. To prove this evaluate both sides for some value of Z, say Z = z 1. LHS = E[X Z = z 1 ]. Now (X Z = z 1 ) is a random variable G defined on the subset A, say, of the sample space where Z(ω) = z 1. If originally a subset B of A had a measure µ(b) now it has the measure µ(b)/µ(a), on this new sample space. The RHS = E[E[(X Z = z 1 ) Y ]] = E[E[G Y ]]. By the law of total probability, E[G] = E[E[G Y ]], so that for Z = z 1, both sides are equal, proving the required result. Another way of looking at this result is as follows. Let us call the event corresponding to Y = y 1, Z = z 1, in the sample space, A(y 1, z 1 ). The event A(, z 1 ), corresponding to Z = z 1 is y i A(y i, z 1 ), the union being disjoint union of pairwise disjoint subsets. Let us call the event corresponding to X = x i, Z = z 1 in the sample space, B(x i, z 1 ). Computing E[X Z = z 1 ] = Σ xi x i p(x i z 1 ) = Σ xi x i p(x i, z 1 )/p(z 1 ), is in 2

words, computing the expectation over each B(x i, z 1 ), and then summing it over all X i. We could compute instead Σ xi Σ yj x i p(x i y j, z 1 )p(y j ) = Σ xi Σ yj x i p(x i, y j, z 1 )/p(z 1 ). Here we are breaking B(x i, z 1 ) into the family of disjoint subsets B(x i, z 1 ) A(y j, z 1 ), computing the expectation over each of these smaller sets according to probability p(x i y j, z 1 ), multiplying by p(y j ) and then summing the values. This latter computation corresponds to the RHS of the equation, E[X Z = z 1 ] = E[E[X Y, Z = z 1 ] Z = z 1 ]. 2.1 Cautionary Remark Martingale theory is about expectations. For reasons of easy readability we will not state conditions such as if X is integrable. Such a condition essentially means E[ X ] <. All the results, in these notes on martingales, may be taken to be valid, unless otherwise stated, only when the expectation operation yields a finite value on f(x 1,, X n ), where X i are the random variables and f( ), the function under consideration. 3 Definition and elementary properties of martingales Let S X 0,, X n, be a sequence of random variables. A second sequence of random variables M M 0,, M n, is said to be a martingale with respect to S, if E[M n+1 X 0,, X n ] = M n. We will replace the sequence X 0,, X n, by F n for short so that the definition reads E[M n+1 F n ] = M n. 1. (Expectation remains invariant with time) Applying expectation operation on both sides of the defining equation of a martingale and using the law of total probability, we have E[M n+1 ] = E[E[M n+1 F n ]] = E[M n ]. 2. (In the martingale definition M n+k can replace M n+1 ) Next, we note that F n+k = F n, X n+1,, X n+k.using the second form of the law of total probability, we have E[M n+k F n ] = E[E[M n+k F n+1 ] F n ]. Now E[M m+r F m ] = M m, r = 1. Suppose it is true that E[M m+r F m ] = M m, r < k, m, then we must have E[M n+k F n+1 ] = M n+1. We therefore have, E[M n+k F n ] = E[E[M n+k F n+1 ] F n ] = E[M n+1 F n ] = M n. Further, by the law of total probability E[M n+k ] = E[E[M n+k F n ]] = E[M n ]. 3. (Orthogonality of martingale increments) Let j k l. We then have E[M j (M l M k )] = E[E[M j (M l M k )] F j ]] = E[M j E[(M l M k ) F j ]] = E[M j (M j M j )] = 0. 3

Next let i j k l. We then have E[(M j M i )(M l M k )] = E[E[(M j M i )(M l M k )] F j ]] = E[(M j M i )E[(M l M k ) F j ]] = E[(M j M i )(M j M j )] = 0. Note that this happens even though (M j M i ), (M l M k ), are not independent random variables. 4. (Variance consequence of orthogonality) V ar[m m+k M m ] = E[(M m+k M m ) 2 ] (E[M m+k M m ]) 2. E[(M m+k M m ) 2 ] = E[(M m+k ) 2 + (M m ) 2 ] 2E[(M m+k )(M m )] = E[(M m+k ) 2 + (M m ) 2 ] 2E[(M m )(M m )] = E[(M m+k ) 2 ] + E[(M m ) 2 ] 2E[(M m ) 2 ] = E[(M m+k ) 2 ] E[(M m ) 2 ]. (E[M m+k M m ]) 2 = (E[M m+k ] E[M m ]) 2 = (E[M m+k ]) 2 + (E[M m ]) 2 2(E[M m+k ])(E[M m ]) = (E[M m+k ]) 2 + (E[M m ]) 2 2(E[M m ])(E[M m ]) = (E[M m+k ]) 2 (E[M m ]) 2. So V ar[m m+k M m ] = V ar[m m+k ] V ar[m m ]. Next since V ar[m 2 M 1 ] = V ar[m 2 ] V ar[m 1 ], it follows that V ar[m 2 ] = V ar[m 1 ]+V ar[m 2 M 1 ]. Similarly, V ar[m n ] = V ar[m 1 ] + Σ n 2 V ar[m i M i 1 ]. 4 Examples of Martingales (a) (Generalized random walk) Let S X 0,, X n, be a sequence of independent random variables with common mean µ. Let M n Σ n 0 X i nµ. As before, let F n denote X 0,, X n. We then have, E[M n+1 F n ] = E[M n + X n+1 µ F n ] = M n + E[X n+1 µ F n ] = M n, making M n a martingale with respect to S. (b) (Products of independent random variables) Let S X 0,, X n, be a sequence of independent random variables with common mean µ, with F n defined as before. Let M n (µ) n Π n 0 X i. Then E[M n+1 F n ] = E[(µ) (n+1) Π n+1 0 X i F n ] = (E[(µ) 1 X n+1 F n ])(µ) n Π n 0 X i = (µ) n Π n 0 X i = M n. 4

(c) (Branching processes) We know that in the case of these processes we have µx n = E[X n+1 X n ] = E[X n+1 F n ]. If we take M n µ n X n, then we find E[M n+1 F n ] = E[µ (n+1) X n+1 X n ] = µ n X n = M n. (d) (Martingales for Markov chains) Suppose S X 0,, X n, is a Markov chain. One way of associating a fair game with the Markov chain is to bet on the next state when the chain is at a state i. The transition matrix would be available to the bettor. The pay off would be 1/p(i, j), p(i, j) > 0, if one bets 1 dollar on j. The expected gain would be 0, since the game is fair. One could define a martingale M n for this situation as a function f(i, n) of the present state i and the time n, i.e., M n f(i, n). In order that M n becomes a martingale we need E[M n+1 F n ] = M n. Now E[M n+1 F n ] = Σ j p(i, j)f(j, n + 1), and M n = f(i, n). Therefore we must have f(i, n) = Σ j p(i, j)f(j, n + 1), as a necessary condition for M n f(i, n) to be a martingale. On the other hand if the function f(i, n) satisfies f(i, n) = Σ j p(i, j)f(j, n + 1), the above argument shows that M n f(i, n) is a martingale. Thus the condition f(i, n) = Σ j p(i, j)f(j, n + 1), ( ) is necessary and sufficient for M n f(i, n) to be a martingale on the Markov chain. We give below a couple of instances of such functions. i. Consider the gamber s ruin Markov chain with p(i, i+1) = p, p(i, i 1) = 1 p. It can be verified that f(i, n) ((1 p)/p) i satisfies the condition (*) above so that M n f(i, n) is a martingale. ii. When p = 1/2 in the gambler s ruin example another possible f(i, n) is the function i 2 n. 5 Optional stopping theorem We state this important result without proof. First we define stopping time for a stochastic process S X 0,, X n,. The positive integer valued, possibly infinite, random variable N, is said to be a random time for S, if the event {N = n} is determined by the random variables X 0,, X n, i.e., if we know the values of X 0,, X n, we can say whether {N = n} is true or not. If P r{n < } = 1, then the random time N is said to be a stopping time. Theorem 5.1. (Martingale Stopping Theorem) Let S X 0,, X n, and let M n be a a martingale for S. Let N be a random time for the process S. If N is bounded or E[N] < and there is a finite M such that E[ M n+1 M n F] < M, then E[M N ] = E[M 0 ]. 5