Stochastic Processes

Similar documents
Stochastic Processes

LECTURE 12 UNIT ROOT, WEAK CONVERGENCE, FUNCTIONAL CLT

1 Independent increments

Stochastic Processes Calcul stochastique. Geneviève Gauthier. HEC Montréal. Stochastic processes. Stochastic processes.

ELEMENTS OF PROBABILITY THEORY

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3

Lecture 17 Brownian motion as a Markov process

Lecture 12. F o s, (1.1) F t := s>t

Lecture 21 Representations of Martingales

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

Stochastic Processes. Winter Term Paolo Di Tella Technische Universität Dresden Institut für Stochastik

Probability and Random Processes

Jump Processes. Richard F. Bass

Let (Ω, F) be a measureable space. A filtration in discrete time is a sequence of. F s F t

Introduction and Preliminaries

1 Which sets have volume 0?

Stochastic integral. Introduction. Ito integral. References. Appendices Stochastic Calculus I. Geneviève Gauthier.

Brownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539

Three hours THE UNIVERSITY OF MANCHESTER. 24th January

18.175: Lecture 3 Integration

An essay on the general theory of stochastic processes

Real Analysis: Homework # 12 Fall Professor: Sinan Gunturk Fall Term 2008

Fundamental Inequalities, Convergence and the Optional Stopping Theorem for Continuous-Time Martingales

Lecture 19 L 2 -Stochastic integration

Probability and Measure

THEOREMS, ETC., FOR MATH 515

MATHS 730 FC Lecture Notes March 5, Introduction

In terms of measures: Exercise 1. Existence of a Gaussian process: Theorem 2. Remark 3.

EDRP lecture 7. Poisson process. Pawe J. Szab owski

Measure-theoretic probability

6. Brownian Motion. Q(A) = P [ ω : x(, ω) A )

MATH 6605: SUMMARY LECTURE NOTES

Verona Course April Lecture 1. Review of probability

A Concise Course on Stochastic Partial Differential Equations

MA 8101 Stokastiske metoder i systemteori

ANALYSIS QUALIFYING EXAM FALL 2016: SOLUTIONS. = lim. F n

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

HOMEWORK ASSIGNMENT 6

Filtrations, Markov Processes and Martingales. Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition

I. ANALYSIS; PROBABILITY

STAT 7032 Probability Spring Wlodek Bryc

Elementary Probability. Exam Number 38119

The strictly 1/2-stable example

1 Continuation of solutions

Stochastic Processes II/ Wahrscheinlichkeitstheorie III. Lecture Notes

(U) =, if 0 U, 1 U, (U) = X, if 0 U, and 1 U. (U) = E, if 0 U, but 1 U. (U) = X \ E if 0 U, but 1 U. n=1 A n, then A M.

Combinatorics in Banach space theory Lecture 12

A Short Introduction to Diffusion Processes and Ito Calculus

Lebesgue Integration: A non-rigorous introduction. What is wrong with Riemann integration?

Lecture 2. We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales.

4 Sums of Independent Random Variables

LECTURE 2: LOCAL TIME FOR BROWNIAN MOTION

MATH MEASURE THEORY AND FOURIER ANALYSIS. Contents

Notes on Measure, Probability and Stochastic Processes. João Lopes Dias

P (A G) dp G P (A G)

1* (10 pts) Let X be a random variable with P (X = 1) = P (X = 1) = 1 2

Real Analysis Problems

Exercises. T 2T. e ita φ(t)dt.

Lecture 6: Contraction mapping, inverse and implicit function theorems

Markov processes Course note 2. Martingale problems, recurrence properties of discrete time chains.

Snell envelope and the pricing of American-style contingent claims

1 Math 241A-B Homework Problem List for F2015 and W2016

2 (Bonus). Let A X consist of points (x, y) such that either x or y is a rational number. Is A measurable? What is its Lebesgue measure?

3 Integration and Expectation

POINCARÉ S INEQUALITY AND DIFFUSIVE EVOLUTION EQUATIONS

Probability and Random Processes

LECTURE 15: COMPLETENESS AND CONVEXITY

A D VA N C E D P R O B A B I L - I T Y

Homework #6 : final examination Due on March 22nd : individual work

Random Process Lecture 1. Fundamentals of Probability

9 Brownian Motion: Construction

Part II Probability and Measure

Applications of Ito s Formula

FACULTY OF SCIENCE FINAL EXAMINATION MATHEMATICS MATH 355. Analysis 4. Examiner: Professor S. W. Drury Date: Tuesday, April 18, 2006 INSTRUCTIONS

Doléans measures. Appendix C. C.1 Introduction

Brownian Motion and Conditional Probability

1. Stochastic Processes and filtrations

Poisson Approximation for Structure Floors

REAL AND COMPLEX ANALYSIS

2 Lebesgue integration

FACULTY OF SCIENCE FINAL EXAMINATION MATHEMATICS MATH 355. Analysis 4. Examiner: Professor S. W. Drury Date: Wednesday, April 18, 2007 INSTRUCTIONS

JUHA KINNUNEN. Harmonic Analysis

Martingale Problems. Abhay G. Bhatt Theoretical Statistics and Mathematics Unit Indian Statistical Institute, Delhi

Stochastic Convergence, Delta Method & Moment Estimators

Stochastic integration. P.J.C. Spreij

Chapter 4. The dominated convergence theorem and applications

Time is discrete and indexed by t =0; 1;:::;T,whereT<1. An individual is interested in maximizing an objective function given by. tu(x t ;a t ); (0.

Widely applicable periodicity results for higher order di erence equations

Discrete-Time Market Models

Recall that if X is a compact metric space, C(X), the space of continuous (real-valued) functions on X, is a Banach space with the norm

R N Completeness and Compactness 1

Solutions to Tutorial 1 (Week 2)

Definition: Lévy Process. Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 2: Lévy Processes. Theorem

( f ^ M _ M 0 )dµ (5.1)

7 Poisson random measures

Heat kernels of some Schrödinger operators

LECTURE NOTES. Introduction to Probability Theory and Stochastic Processes (STATS)

Analysis Finite and Infinite Sets The Real Numbers The Cantor Set

Euler 2D and coupled systems: Coherent structures, solutions and stable measures

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

Transcription:

Stochastic Processes A very simple introduction Péter Medvegyev 2009, January Medvegyev (CEU) Stochastic Processes 2009, January 1 / 54

Summary from measure theory De nition (X, A) is a measurable space if A 6=? is a set of subsets of X such that 1 if A 2 A, then A c 2 A, 2 if (A n ) is a countable sequence of sets of A, then [ n A n 2 A, 3 if (A n ) is a countable sequence of sets of A, then \ n A n 2 A. The set of sets A is called σ-algebra, the sets in A are the measurable sets. Medvegyev (CEU) Stochastic Processes 2009, January 2 / 54

Measurable functions and random variables De nition If (X, A) and (Y, B) are measurable spaces a mapping f : X! Y is called measurable if for every B 2 B f 1 (B) $ fx 2 X j f (x) 2 Bg 2 A. If Y = R then B is always the set of Borel measurable sets which is the smallest σ-algebra containing the intervals. If the image space is the set of real numbers we are talking about measurable functions. If (Ω, A, P) is a probability space then the random variables are the measurable (extended real valued) functions. We say that two random variables ξ 1 and ξ 2 are equivalent or indistinguishable if P (ξ 1 6= ξ 2 ) = 0. Medvegyev (CEU) Stochastic Processes 2009, January 3 / 54

Measurable functions and random variables De nition If s = k c k χ Ak then s is a step function. If the sets A k are measurable then s is a measurable step function. Theorem If f 0 is a measurable function then there is a sequence of measurable step functions (s n ) such that 0 s n % f. Medvegyev (CEU) Stochastic Processes 2009, January 4 / 54

Measurable functions and random variables Theorem Let (X, A) be a measurable space. 1 If f and g are measurable functions then f + g, f g, f /g are measurable. (Of course only when they are well-de ned.) If λ is a real number and f is a measurable function then λf is also measurable. 2 If (f n ) is a sequence of measurable functions then are measurable. sup f n, inf n n f n, lim supf n, lim inf f n n! n! Medvegyev (CEU) Stochastic Processes 2009, January 5 / 54

Measurable functions and random variables Theorem If (f n ) is a sequence of measurable functions and the limit exists on a set A then is also measurable. g (x) $ f (x) $ lim n! f n (x) f (x) if x 2 A 0 if x /2 A Medvegyev (CEU) Stochastic Processes 2009, January 6 / 54

Integral and expected value De nition The integral of a step function s 0 is Z sdµ $ c k µ (A k ). X k The integral of a measurable function f 0 is Z Z fdµ $ sup sdµ X 0sf X where s 0 is a measurable step function. If f is arbitrary then Z Z Z fdµ $ f + dµ f dµ assuming that the sum is not of the form integrable if the integral is nite. X X X. A function f is Medvegyev (CEU) Stochastic Processes 2009, January 7 / 54

Integral and expected value De nition If µ is a probability measure then Z E (ξ) $ Ω ξdp. Medvegyev (CEU) Stochastic Processes 2009, January 8 / 54

Integral and expected value Theorem If f and g are measurable functions then Z Z Z (f + g) dµ = fdµ + gdµ X X X assuming that all the three integrals exists. If two exist then one can prove that the third also exits. Medvegyev (CEU) Stochastic Processes 2009, January 9 / 54

Integral and expected value Theorem (Monotone convergence) If 0 f n % f are measurable functions then Z Z f n dµ % fdµ Theorem (Dominated convergence) X If (f n ) are measurable functions and f n! f and there is an integrable function g such that jf n j g the Z X Z f n dµ! X X fdµ and Z X jf n f j dµ! 0 Medvegyev (CEU) Stochastic Processes 2009, January 10 / 54

Banach space of integrable functions De nition If p 1 and jf j p is integrable then we say that f 2 L p. Theorem The equivalence classes of integrable functions L p with the norm r Z kf k p $ p jf j p dµ X is a Banach space, that is a complete normed space. Medvegyev (CEU) Stochastic Processes 2009, January 11 / 54

Conditional expectation De nition Let ξ be a random variable F A is a sub σ-algebra. Assume that E (ξ) exists. The conditional expectation E (ξ j F) is an F-measurable random variable such that Z Z ξdp = E (ξ j F) dp, 8F 2 F. F F P (A j F) $ E (χ A j F). Example If F = f, Ωg then E (ξ j F) = E (ξ). Medvegyev (CEU) Stochastic Processes 2009, January 12 / 54

Theorem Assume that all variables below are integrable. If ξ and F are independent, then E (ξ j F) = E (ξ). Specially if F = f, Ωg, then E (ξ j F) = E (ξ). If ξ is F-measurable and the conditional expectation E (ξ j F) exists, then E (ξ j F) = ξ. If ξ η, then E (ξ j F) E (η j F),specially 0 P (A j F) 1, je (ξ j F)j E (jξj j F). E (ξ + η j F) = E (ξ j F) + E (η j F). If A and B are disjoint sets, then P (A [ B j F) = P (A j F) + P (B j F), P (B n A j F) = P (B j F) P (A j F), A B. Medvegyev (CEU) Stochastic Processes 2009, January 13 / 54

Tower laws Theorem If G F, then E (E (ξ j F) j G) = E (ξ j G). Specially If G F, then E (E (ξ j F)) = E (ξ). E (E (ξ j G) j F) = E (ξ j G). Medvegyev (CEU) Stochastic Processes 2009, January 14 / 54

Monotone convergence Theorem The monotone convergence theorem is satis ed, that is, if ξ n 0, and ξ n % ξ,then E (ξ n j F). % E (ξ j F). Specially, if A n % A, then If ξ n 0, then P (A n j F). % P (A j F). E (ξ n j F) = E n=1! ξ n j F. n=1 If the sets (A n ) are disjoint and A = [ n A n, then P (A n j F) = P (A j F). n=1 Medvegyev (CEU) Stochastic Processes 2009, January 15 / 54

Dominated convergence Theorem The dominated convergence theorem is satis ed, that is if where E (η) <, then lim ξ n! n = ξ, and jξ nj η, lim E (jξ n! n ξj j F) = 0, lim E (ξ n! n j F) = E (ξ j F). Specially, if A n & A, then lim P (A n j F) = P (A j F). n! Medvegyev (CEU) Stochastic Processes 2009, January 16 / 54

Homogeneity Theorem If variable η is F-measurable, and E (ξ) and E (η ξ) are nite, then E (η ξ j F) = η E (ξ j F). It is also true, if η and ξ are nonnegative. Medvegyev (CEU) Stochastic Processes 2009, January 17 / 54

De nition of Stochastic Processes Let us x a probability space (Ω, A, P). As in probability theory we refer to the set of real-valued (Ω, A)-measurable functions as random variables. In the theory of stochastic processes random variables very often have in nite value. Hence the image space of the measurable functions is not R but the set of extended real numbers R $ [, ]. The most important examples of random variables with in nite value are stopping times. Stopping times give the random time of the occurrence of observable events. If for a certain outcome ω the event never occurs it is reasonable to say that the value of the stopping time for this ω is +. In the most general sense stochastic processes are such functions X (t, ω) that for any xed parameter t the mappings ω 7! X (t, ω) are random variables on (Ω, A, P). Medvegyev (CEU) Stochastic Processes 2009, January 18 / 54

De nition of Stochastic Processes The set of possible time parameters Θ is some subset of the extended real numbers. In the theory of continuous-time stochastic processes Θ is an interval, generally Θ = R + $ [0, ), but sometimes Θ = [0, ] and Θ = (0, ) is also possible. If we do not say explicitly what the domain of de nition of the stochastic process is, then Θ is R +. Medvegyev (CEU) Stochastic Processes 2009, January 19 / 54

De nition of Stochastic Processes It is very important to append some remarks to this de nition. In probability theory the random variables are equivalence classes which means that the random variables X (t) are de ned up to measure zero sets. This means that in general X (t, ω) is meaningless for a xed ω. If the possible values of the time parameter t are countable then we can select from the equivalence classes X (t) one element and we can x a measure zero set and outside of this set the expressions X (t, ω) are meaningful. But this is impossible if Θ is not countable. Therefore we shall always assume that X (t) is a function already carefully selected from its equivalence class. To put it in another way: when one de nes a stochastic process, one should x the space of possible trajectories and the stochastic processes are function-valued random variables which are de ned on the space (Ω, A, P). Medvegyev (CEU) Stochastic Processes 2009, January 20 / 54

De nition of Stochastic Processes De nition Let us x the probability space (Ω, A, P) and the set of possible time parameters Θ. The function X de ned on Θ Ω is a stochastic process over Θ Ω if for every t 2 Θ it is measurable on (Ω, A, P) in its second variable. De nition If we x an outcome ω 2 Ω then the function t 7! X (t, ω) de ned over Θ is the trajectory or realization of X corresponding to the outcome ω. If all the trajectories of the process X have a certain property then we say that the process itself has this property. For example if all the trajectories of X are continuous then we say that X is continuous, if all the trajectories of X have nite variation then we say that X has nite variation, etc. Medvegyev (CEU) Stochastic Processes 2009, January 21 / 54

When do stochastic process be equal? De nition Let X and Y be two stochastic processes on the probability space (Ω, A, P). 1 The process X is a modi cation of the process Y if for all t 2 Θ the variables X (t) and Y (t) are almost surely equal, that is for all t 2 Θ P (X (t) = Y (t)) $ P (fω : X (t, ω) = Y (t, ω)g) = 1. By this de nition, the set of outcomes ω where X (t, ω) 6= Y (t, ω), can depend on t 2 Θ. 2 The processes X and Y are indistinguishable if there is a set N Ω which has probability zero, and whenever ω /2 N then X (ω) = Y (ω), that is X (t, ω) = Y (t, ω) for all t 2 Θ and ω /2 N. Medvegyev (CEU) Stochastic Processes 2009, January 22 / 54

When do stochastic process be equal? De nition Let X and Y be two stochastic processes on the probability space (Ω, A, P). X = Y means that X and Y are indistinguishable, that is the trajectories are almost surely equal. Medvegyev (CEU) Stochastic Processes 2009, January 23 / 54

Filtration, usual conditions De nition If (F t ) t is a monotone increasing set of σ-algebras, then F = (F t ) t is called ltration, that is F is a ltration if for s < t F s F t. The stochastic process X is called adapted, if X (t) is F t measurable for all t. De nition The ltration (Ω, A, P, F)satis es the usual conditions if 1 A is complete, that is if N M, and P (M) = 0, then N 2 A. 2 F t for all t contains the measure zero sets, 3 the ltration F is right-continuous, that is for all t F t = F t+ $ \ s>t F s. We will always assume, that the usual conditions are satis ed. We will also assume that every process is adapted to the given ltration. Medvegyev (CEU) Stochastic Processes 2009, January 24 / 54

Filtration, usual conditions The usual interpretation of the σ-algebra F t is that it contains the events which occurred up to time t, that is F t contains the information which is available at moment t. As F t is the information at moment t one can interpret F t+ as the information available in nitesimally just after t. If a process has "speed", then one can foresee the future in nitesimally as by de nition in this case the right and left derivatives are equal and one knows the left derivative at time t. The stochastic processes are very often non-di erentiable. Even in this case one should assume this "in nitesimal wisdom". The trick is that one should add the measure zero sets to the ltration anyway, and surprisingly it makes the ltrations right-continuous. We will return to this later. Medvegyev (CEU) Stochastic Processes 2009, January 25 / 54

Generated ltration It is a quite natural question how one can de ne a ltration F? De nition Let X be a stochastic process that is let X be a function of two variables..let us de ne the σ-algebras Ft X A generated by the sets fx (t 1 ) 2 I 1,..., X (t n ) 2 I n g where t 1,..., t n t are arbitrary elements in Θ and I 1,..., I n are arbitrary intervals. Obviously if s < t then Fs X Ft X, hence F X is really a ltration. F X is called the ltration generated by X. There is no guarantee that F X satis es the usual conditions. Very often if we add to F X the measure zero sets the new ltration is right-continuous and the usual conditions hold. Medvegyev (CEU) Stochastic Processes 2009, January 26 / 54

Martingales De nition The stochastic process X is a martingale relative to a ltration F, if 1 E (jx (t)j) <, for all t, 2 if s < t, then E (X (t) j F s ) = X (s), 3 the trajectories are right-continuous, and have limits from the left. De nition The L 2 -bounded martingales are denoted by H 2. It is trivial from the assumptions that every martingale is adapted. Medvegyev (CEU) Stochastic Processes 2009, January 27 / 54

Wiener processes De nition The process w called Wiener process, if 1 w (0) = 0, 2 w has stationary and independent increments, 3 w has continuous trajectories, 4 w (t) ' N (0, t) (= N 0, σ 2 ). 5 w is adapted. Medvegyev (CEU) Stochastic Processes 2009, January 28 / 54

Wiener processes on nite intervals Example On any nite interval Wiener processes are in H 2, on in nite intervals the Wiener processes are just martingales. Only the second condition is not trivial. E (w (t) j F s ) = E (w (s) + w (t) w (s) j F s ) = = E (w (s) j F s ) + E (w (t) w (s) j F s ) = = w (s) + E (w (t) w (s)) = w (s). Medvegyev (CEU) Stochastic Processes 2009, January 29 / 54

Generated ltration of Wiener processes Let A $ fω : w (ω) is zero for some interval [0, δ (ω)]g. As F0 w = f, Ωg obviously A /2 F 0 w. On the other hand A 2 F w 1/n for every n > 0 as for every n > 0 we know that A holds or not. Hence A 2 \ n F1/n w = F 0+ w. Hence F w is not right-continuous. One can prove that if F t $ σ (F w t, N ) where N is the set of events with probability zero, then F satis es the usual conditions. Very often F is generated by some nite number of Wiener processes, that is F t $ σ (F w 1 t,..., F w n t, N ). Medvegyev (CEU) Stochastic Processes 2009, January 30 / 54

Poisson processes De nition The process π called Poisson process, if 1 π (0) = 0, 2 the trajectories are continuous from the right and have limits from the left and have pure jumps of height one, 3 π (t) ' P (λt), that is P (π (t) = n) = (λt)n n! exp ( λt). Medvegyev (CEU) Stochastic Processes 2009, January 31 / 54

Lévy processes De nition An adapted process X is called Lévy process (with respect to F) if 1 X (0) = 0, 2 X has independent increments with respect to F, that is if t > s then X (t) X (s) is independent of F s, 3 X has stationary increments, 4 the trajectories are continuous from the right and have limits from the left. Medvegyev (CEU) Stochastic Processes 2009, January 32 / 54

Martingales and Lévy processes Example If the increments of a Lévy process X has zero expected value, then X is a martingale. E (X (t) j F s ) = E (X (s) + X (t) X (s) j F s ) = = E (X (s) j F s ) + E (X (t) X (s) j F s ) = = X (s) + E (X (t) X (s)) = X (s). Medvegyev (CEU) Stochastic Processes 2009, January 33 / 54

Independence and Fourier transformation De nition If ξ is a vector of random variables then ϕ (u) $ E (exp (i hu, ξi)) is called the Fourier transform or characteristic function of vector ξ and L (u) $ E (exp (hu, ξi)) is called the moment generating function of ξ. Theorem The Fourier transform determines the distribution of ξ. Theorem ξ and η are independent if and only if ϕ (ξ,η) (u, v) = E (exp (iuξ + ivη)) = E (exp (iuξ)) E (exp (ivη)) = = ϕ ξ (u) ϕ η (v). Medvegyev (CEU) Stochastic Processes 2009, January 34 / 54

Independence and Fourier transformation The easy part is:that if ξ and η are independent then ϕ (ξ,η) (u, v) = ϕ ξ (u) ϕ η (v). By de nition ξ and η are independent if the generated σ-algebras are independent, therefore exp (iuξ) and exp (iv η) are independent and the independent variables are uncorrelated. Medvegyev (CEU) Stochastic Processes 2009, January 35 / 54

Remarks 1 The moment generating function generally does not determine the distribution. Medvegyev (CEU) Stochastic Processes 2009, January 36 / 54

Remarks 1 The moment generating function generally does not determine the distribution. 2 ϕ (u) is well-de ned and nite for every u, but L (u) is not necessarily nite for every u. Medvegyev (CEU) Stochastic Processes 2009, January 36 / 54

Remarks 1 The moment generating function generally does not determine the distribution. 2 ϕ (u) is well-de ned and nite for every u, but L (u) is not necessarily nite for every u. 3 If L is well-de ned and nite in a neighborhood of the origin, then L determines the distribution. Medvegyev (CEU) Stochastic Processes 2009, January 36 / 54

Remarks 1 The moment generating function generally does not determine the distribution. 2 ϕ (u) is well-de ned and nite for every u, but L (u) is not necessarily nite for every u. 3 If L is well-de ned and nite in a neighborhood of the origin, then L determines the distribution. 4 If ξ 0 and L (s) $ E (exp ( sξ)), s 0 then L determines the distribution. (In this case L is the Laplace transform of ξ.) Medvegyev (CEU) Stochastic Processes 2009, January 36 / 54

Exponential martingales Example If X is a Lévy process, then for every u M (t, u) $ exp (iux (t)) ϕ t (u) is a martingale, where ϕ t (u) $ E (exp (iu (X (t)))). Alternatively if L t (u) exists for some u then one can de ne M (t, u) $ exp (ux (t)). L t (u) Medvegyev (CEU) Stochastic Processes 2009, January 37 / 54

Exponential martingales exp (iux (t)) E (M (t, u) j F s ) = E j F s = ϕ t (u) exp (iu [X (s) + X (t) = E ϕ t (u) exp (iu [X (s) + X (t) = E = E M (s, u) = = X (s)]) j F s = X (s)]) j F s = ϕ s (u) ϕ t s (u) exp (iu [X (t) X (s)]) j F s = ϕ t s (u) M (s, u) ϕ t s (u) E (exp (iu [X (t) X (s)]) j F s ) = M (s, u) E (exp (iu [X (t) X (s)])) = M (s, u). ϕ t s (u) Medvegyev (CEU) Stochastic Processes 2009, January 38 / 54

Exponential martingale of Wiener processes Example If w is a Wiener process, then for every u exp (iuw (t)) M (t, u) $ iuw exp ( tu 2 /2) = exp (t) + tu2 2 tu M (t, u) $ 2 exp uw (t) 2 are martingales, called, the exponential martingales of w. Medvegyev (CEU) Stochastic Processes 2009, January 39 / 54

Exponential martingale of Wiener processes The second one is easy E (exp (uw (t))) = E L N (0,1) (u) = = = exp u p tn (0, 1) = L N (0,1) u p t. Z 1 1 p exp 2π 2 x 2 exp (ux) dx = Z 1 x 2 2ux p exp dx = 2π 2 1 p 2π Z exp u 2 = exp 2 x 2 2ux + u 2 + u2 dx = 2 2! (x u) 2 u 2 dx = exp 2 2 1 p 2π Z exp. Medvegyev (CEU) Stochastic Processes 2009, January 40 / 54

Exponential martingale of Wiener processes Also E (exp (iuw (t))) = E exp iu p tn (0, 1) = ϕ N (0,1) u p t. ϕ N (0,1) (u) = = = Z 1 1 p exp 2π 2 x 2 exp (iux) dx = Z 1 1 p exp 2π 2 x 2 (cos ux + i sin ux) dx = Z 1 1 p exp 2π 2 x 2 cos uxdx. Medvegyev (CEU) Stochastic Processes 2009, January 41 / 54

Exponential martingale of Wiener processes Di erentiating and integrating by parts = d du ϕ N (0,1) (u) = p 1 Z exp 2π = p 1 Z 2π 1 exp 2 x 2 sin ux = 0 ( x) exp Z 1 p u exp 2π 1 p 2π u 1 2 x 2 sin uxdx = 1 2 x 2 Z d cos uxdx = du 1 exp 2 x 2 cos uxdx = 1 2 x 2 cos uxdx = uϕ (u). As ϕ (0) = 1 the solution of this di erential equation is ϕ (u) = exp u 2 /2. Medvegyev (CEU) Stochastic Processes 2009, January 42 / 54

Compensated Poisson processes Example The compensated Poisson process π (t) λt is a martingale. π (t) λt is a Lévy process with expected value zero. Medvegyev (CEU) Stochastic Processes 2009, January 43 / 54

Compensated square of Wiener processes Example If w is Wiener process, then N (t) $ w 2 (t) t is a martingale. E (N (t) j F s ) = E w 2 (t) t j F s = = E (w (s) + w (t) w (s)) 2 t j F s. E (w (t) w (s)) 2 j F s = E (w (t) w (s)) 2 = = E N 2 (0, t s) = t s. E (w (s) (w (t) w (s)) j F s ) = w (s) E ((w (t) w (s)) j F s ) = 0. Medvegyev (CEU) Stochastic Processes 2009, January 44 / 54

Stopping times De nition A random variable 0 τ is called stopping time, if fτ tg 2 F t for all t This is the same as τ ^ t is F t -measurable for every t that is τ ^ t is "observable" for every t. Theorem If the trajectories of the adapted stochastic process X are right- or left-continuous, or the process is measurable with respect to the σ-algebra generated by the class of the adapted right-continuous processes, then for any Borel set B the hitting time is a stopping time. τ $ inf ft : X (t) 2 Bg Medvegyev (CEU) Stochastic Processes 2009, January 45 / 54

Stopped process, stopped variable De nition Let X be a stochastic process, and let τ be a stopping time. 1 By a stopped or truncated process we mean the process 2 We shall call the random variable a stopped variable. X τ (t, ω) $ X (τ (ω) ^ t, ω). X τ (ω) $ X (τ (ω), ω) 3 The stopped σ-algebra F τ is the set of events A 2 A for which for all t A \ fτ tg 2 F t. Medvegyev (CEU) Stochastic Processes 2009, January 46 / 54

Stopped process, stopped variable Instead of X τ we shall very often use the more readable notation X (τ). Observe that the de nition of stopped variable is not entirely correct as X is generally not de ned on the set fτ = g and it is not clear what the de nition of X τ on this set is. If τ (ω) /2 Θ then one can use the de nition X τ (ω) $ 0. If one uses the convention that the product of an unde ned value with zero is zero, then one can write the de nition of the stopped variable X τ in the following way: X τ (ω) $ X (τ (ω), ω) χ (τ 2 Θ) (ω). Medvegyev (CEU) Stochastic Processes 2009, January 47 / 54

Optional Sampling Theorem Theorem If M is a martingale and τ is a stopping time, then the stopped process M τ is also a martingale. Theorem An adapted process M is a martingale if and only if, the trajectories of M are right-continuous and they have limits from the left and for any bounded stopping time τ E (M τ ) = E (M (0)), where as above M τ (ω) $ M (τ (ω), ω). Medvegyev (CEU) Stochastic Processes 2009, January 48 / 54

Uniformly integrable martingales When can one drop the condition that τ is bounded? For any T obviously τ ^ T is bounded so E (M (τ ^ T )) = E (M (0)). Can we take the limit T! under the integral sign: E (M (τ)) = E lim M (τ ^ T ) T! = lim E (M (τ ^ T )) = E (M (0))? T! If τ < and M has an integrable majorant, that is jm (t)j ξ 2 L 1 (Ω) then the answer is certainly yes. There is a generalization. There is a special concept, called "uniformly integrable" martingale which generalizes the mentioned situation. An important case when M is uniformly integrable but there is no integrable majorant is when M 2 H 2. Medvegyev (CEU) Stochastic Processes 2009, January 49 / 54

Uniformly integrable martingales Example If w is a Wiener process and τ a $ inf (t : w (t) a) then as w is continuous and as w is unbounded τ a < and w (τ a ) = a. Hence if a 6= 0 Observe that E (w (τ a )) = E (a) = a 6= E (w (0)) = 0. w ( ) $ lim t! w (t) is not well-de ned. Observe that w is in H 2 on any nite interval but w /2 H 2 on R +. Medvegyev (CEU) Stochastic Processes 2009, January 50 / 54

Optional Sampling Theorem for uniformly integrable martingales Theorem If M is a uniformly integrable martingale, then 1 one can extend M from [0, ) to [0, ], 2 the extended process remains a martingale and 3 E (M (τ)) = E (M (0)) holds for any stopping time τ. Medvegyev (CEU) Stochastic Processes 2009, January 51 / 54

Ruin probabilities Example If a < 0 < b and τ a and τ b are the respective rst passage times of some Wiener process w, then P (τ a < τ b ) = b b a, P (τ b < τ a ) = a b a. With probability one the trajectories of w are unbounded. Therefore as w starts from the origin the trajectories of w nally leave the interval [a, b]. So P (τ a < τ b ) + P (τ b < τ a ) = 1. If τ $ τ a ^ τ b then w τ is a bounded martingale. Hence one can use the Optional Sampling Theorem. Obviously w τ τ is ether a or b, hence E (w τ τ ) = ap (τ a < τ b ) + bp (τ b < τ a ) = E (w τ (0)) = 0. We have two equations with two unknowns. Solving this system of linear equations, one can easily deduce the formulas above. Medvegyev (CEU) Stochastic Processes 2009, January 52 / 54

Energy Equality Theorem If M 2 H 2, then E (M (t) M (s)) 2 = E M 2 (t) E M 2 (s) By the properties of the conditional expectation for $ 2 E (M (s) (M (s) M (t))) the di erence of the two sides $ 2 E (M (s) (M (s) M (t))) = = 2 E (M (s) E (M (s) M (t) j F s )) = = 2 E (M (s) 0) = 0. Medvegyev (CEU) Stochastic Processes 2009, January 53 / 54

Energy Equality and the Uniform Integrability Recall that M 2 H 2 if km (t)k 2 K. By the Energy Equality the function t 7! km (t)k is increasing. As it is bounded it is convergent. Hence for any ε > 0 km (t) M (s)k jkm (t)k km (s)kj < ε it t, s N (ε). Hence (M (t)) t is a Cauchy sequence in L 2 (Ω). As L 2 (Ω) is complete it is convergent, that is M (t)! M ( ) and the convergence holds in L 2 (Ω). As M is a martingale E (M (t) j F s ) = M (s). If t!, then as convergence in L 2 (Ω) implies convergence in L 1 (Ω) E (M ( ) j F s ) = E lim M (t) j F s = lim E (M (t) j F s ) = t! t! = lim M (s) = M (s). t! Medvegyev (CEU) Stochastic Processes 2009, January 54 / 54