Stochastic Processes and Stochastic Analysis Notes of the Lecture by Prof. M. Schweizer

Size: px
Start display at page:

Download "Stochastic Processes and Stochastic Analysis Notes of the Lecture by Prof. M. Schweizer"

Transcription

1 Stochastic Processes and Stochastic Analysis Notes of the Lecture by Prof. M. Schweizer Luca Gugelmann Thomas Rast Sommersemester 27 Warning: We are sure there are lots of mistakes in these notes. Use at your own risk! Corrections and other feedback would be greatly appreciated and can be sent to If you report an error please always state what version (the first number on the Id line below) you found it in. For further information see: Revision 1375 March 26, 21 1

2 2 Contents Preliminaries 3 1 Brownian Motion: Definition and Construction Definition and First Remarks Donsker s theorem Some applications Properties of Brownian Motion Transformations and First Properties Path properties of BM Martingale properties of BM The law of the iterated logarithm Markov processes Basic concepts Markov property and strong Markov property Some applications Generators and martingale problems Stochastic calculus Continuous semimartingales and quadratic variation Stochastic integrals Itô s formula The Girsanov transformation The Kunita-Watanabe decomposition Itô s representation theorem Stochastic differential equations: Basic notions Connections to PDEs Lévy processes Basic concepts Some properties of Lévy processes

3 3 Introduction Topics: Brownian motion Markov processes Stochastic analysis Lévy processes Prerequisite: Probability theory, to the extent of the course last semester. See script and lecture notes. Literature will follow, but the course is not based on any book in particular. Preliminaries Goal: collect/recall a few basic notions and facts. Definition. (Ω, F, P ) a probability space. A filtration on (Ω, F) over [, ) is an increasing family F = (F t ) t of σ-fields F t F, i.e. F s F t, s t. Intuition: F t is a family of events observable up to time t, briefly: the information available in t. F is called right-continuous if F t = F t+ := ε> F t+ε, t. F is called (P -)complete if F is (P -)complete (i.e. any subset of a nullset is measurable), and if F contains all P -nullsets of F. F is said to satisfy the usual conditions if it is right-continuous and complete. Definition. A (real-valued) stochastic process X = (X t ) t is any collection of random variables X t : Ω R indexed by t. More generally one can consider an S-valued stochastic process where (S, S) is a measurable space and indexed by some set I (instead of [, )); but to describe a temporal evolution, t is natural. Three ways of thinking about a stochastic process: 1. A collection of random variables X t on (Ω, F), indexed by t. 2. A family of random functions t X t (ω) on [, ) for every ω Ω; X (ω) for fixed ω is called the path or trajectory corresponding to ω. 3. A mapping X : Ω [, ) R on the product space Ω := Ω [, ). In (2), one can also view X as a mapping from Ω to the function space R [, ), ω X (ω), i.e., a function-valued map on Ω. We say that X has property A (e.g. continuous, increasing, bounded,... ) if X (ω) has property A for P -almost all ω. Note.: Quantitative aspects of A may depend on ω; for example, X a bounded stochastic process usually means: with probability 1, X (ω) is bounded by a constant c(ω). If c does not depend on ω, one says bounded uniformly in ω.

4 4 Now think of X as evolution of system over time; how can this be linked to evolution of information? Call X (F-)adapted if for every t, X t : Ω R is F t -measurable (i.e., X t is observable at time t). Now view X as mapping Ω R; X is called product-measurable w.r.t. the product-σ-field F := F B([, )) on Ω = Ω [, ). There are other possible σ-fields on Ω; one is the optional σ-field O which is generated by all processes which are adapted and have RCLL paths (Right Continuous, Left Limits); X is optional if it is O-measurable. Remark.: g : [, ) R is RCLL if (i) lim s t, s>t g(s) = g(t), t (RC). (ii) lim s t, s<t g(s) exists, t > (LL). Useful result: Theorem..1 (Monotone class theorem). Let M be a set of bounded real-valued functions on some set S. Suppose that M is closed under multiplication and set A := σ(m). Let H be a real vector space of bounded real-valued functions on S and assume H has the following properties: (i) H contains M. (ii) H contains the constant function 1. (iii) H is closed under monotone bounded convergence, i.e., if f 1 f 2... are in H and f := lim n f n is again bounded, then f H. Then H contains all bounded A-measurable functions. The same conclusion is true if H is not a vector space, but in addition closed under uniform convergence, and M is in addition closed under addition (i.e., M is an algebra) and contains the constant function 1. Proof. See Dellacherie/Meyer, I.22, and Protter, Th. I.8. Typical use: want to prove property B for all measurable functions. Then take H := {all bdd. f sat. B}; check (i), (ii), (iii); check that functions in M do satisfy B; if M is good, then this is enough. Example (Exercise!).: Prove that every optional process is adapted. Now view X as in (2) as function valued map X : Ω R [, ). Define coordinate maps Y t : R [, ) R y Y t (y) := y(t) and σ-field B(R) [, ) := σ(y t ; t ). (This is much smaller than the productσ-field, see WT Lemma II.3.6.) Then X is a stochastic process iff X is B(R)- measurable; see Kallenberg, Lemma 2.1. The law/distribution of the stochastic process X under P is a probability measure on (R [, ), B(R [, ) )). So for many things, we can think of stochastic process as measures on some function space. How to describe distribution P X 1 more compactly? It is enough to know all finite-dimensional marginal distributions (fdmds), i.e., the probability measures

5 5 µ (I) := P (X t1,..., X tn ) 1 on R n = R I, for all finite I = {t 1,..., t n } [, ). Then the µ (I), I [, ) finite, determine P X 1 on B(R [, ) ) uniquely; see WT, Lemma II.2.2. A stochastic process is Gaussian if all its fdmds are (multivariate) normal distributions. This says nothing about path properties. The distribution of a Gaussian process is uniquely determined by t E[X t and (s, t) Cov(X s, X t ). Finally, two stochastic processes X, X are called versions/modifications of each other if P [X t = X t = 1, t. They are called indistinguishable if P [X t = X t, t = 1. (More precisely: if t {X t X t } has outer P -measure.) The second notion is usually strictly stronger than the first one; the notions agree e.g. if X, X are both RC. 1 Brownian Motion: Definition and Construction 1.1 Definition and First Remarks Let (Ω, F, P ) a probability space. Definition. A Brownian motion (BM) on [, 1 is a stochastic process with: W = (W t ) t 1 (BM1) P [W = = 1: BM starts at P -a.s. (BM2 ) For any n N and any points = t < t 1 < < t n 1, the increments W tk W tk 1, k = 1,..., n, are independent and N (, t k t k 1 ): independent, Gaussian increments. (BM3) P -almost all trajectories W (ω) are continuous on [, 1. Then the image measure/distribution on (C[, 1, B(C[, 1)) is called Wiener measure. Remark.:. Existence of such a process is non-trivial. 1. Wiener measure is determined by its fdmds because B(C[, 1) is generated by coordinate maps; see WT, V.4.1. and Lemma II BM with respect to a filtration F = (F t ) t 1 : F-adapted stochastic process W = (W t ) t 1 with (BM1), (BM3) and (BM2) for s < t, W t W s is independent of F s and N (, t s). Proposition A Brownian Motion can be characterized as the unique Gaussian process X = (X t ) t 1 with P -a.s. continuous trajectories and with E[X t, Cov(X s, X t ) = min(s, t). Proof. (BM1) plus (BM2 ) are equivalent to saying that the fdmds of W are multivariate normal, because (W, W t1,..., W tn ) is an affine transformation of

6 1.1 Definition and First Remarks 6 X n t 1 n 1/n 1 t Figure 1: Rescaled linear interpolation approach to BM (W t1, W t2 W t1,..., W tn W tn 1 ). So a BM W is a Gaussian process and E[W t =, and for s < t Cov(W s, W t ) = Cov(W s, W t W s ) + Var[W s = s = min(s, t). }{{} = by (BM2 ) Conversely, if X is as above, then E[X =, Var[X =, so X = P -a.s.; and any vector of increments is then multivariate normal, as affine transformation, because X is Gaussian. Moreover E[X t X s = and Var[X t X s = Cov(X t X s, X t X s ) = t s s + s = t s; and finally, for s t u v, Cov(X t X s, X v X u ) = t t s + s =, so increments are uncorrelated, hence independent, because they are multivariate normal. So X is BM. Ways of proving existence of BM: 1. Limit of a rescaled random walk: (Y k ) k N i.i.d., E[Y k =, Var[Y k = 1; S n = n k=1 Y k; X n = (Xt n ) t 1 piecewise linear interpolation of Xl/n n = S l / n (Figure 1). Call µ n the distribution of X n on (C[, 1, B(C[, 1)). Then µ n µ weakly as n, so µ is the Wiener measure. This is Donsker s theorem. 2. Construction via fdmds: (BM1) and (BM2 ) define a system of fdmds; Kolmogorov s consistency theorem (WT, Theorem II.3.5) thus gives existence of a probability measure µ on (R [,1, B(R [,1 )) with these fdmds. Then use Kolmogorov s continuity criterion to show that coordinate process Y on R [,1 is uniformly continuous at rational points, µ-a.s. Then define BM by { continuous continuation of Y (ω) µ-a.e. W (ω) := on a µ-nullset Then W is BM under P := µ on Ω := R [,1. For details see Karatzas/Shreve,

7 1.2 Donsker s theorem 7 3. Random superposition of deterministic functions: Take (Z n ) n N i.i.d. N (, 1) and a sequence (φ n ) n N of deterministic functions in C [, 1. Remark.: W t (ω) := Z n (ω)φ n (t), t 1. n=1 Under suitable assumptions on (φ n ), this W is a BM (Lévy s construction). See Steele, Section BM on [, ): take independent copies W k of BM on [, 1 and glue them together: n W t := W1 k + Wt n n+1 if n < t n + 1. k=1 2. BM in R d : simply (W 1,..., W d ) where W i are independent BMs in R. 1.2 Donsker s theorem (Ω, F, P ) probability space; (Y k ) k N i.i.d. with E[Y k =, Var[Y k = 1; S n := n k=1 Y k and piecewise linear interpolation (as in Figure 1 above) X n t = 1 n S [nt + 1 n Y [nt+1 (nt [nt). µ n := P (X n ) 1 M 1 (C[, 1). Then: Theorem (Donsker s theorem). Under the above assumptions, µ n µ as n, where µ is the Wiener measure. Remark.: 1. Existence plus construction of Wiener measure (or BM). 2. Invariance principle: always same limit irrespective of the exact distribution of the Y k s. 3. As in CLT, one can generalize from i.i.d. to a triangular array. 4. Functional version of CLT: for t = 1, X n 1 = 1 n S n yields CLT; here, we look at the whole process (all t simultaneously). 5. If F : C[, 1 S is continuous, then X n X implies that F (X n ) F (X). One application in finance: convergence (in distribution) of multiplicative binomial models (Cox-Ross-Rubinstein) to geometric BM (Black- Scholes model). To prove Theorem 1.2.1, go back to WT, Theorem V.4.3.: need to prove a) weak convergence of all fdmds (via CLT) b) tightness of (µ n ) n N (real work). Preparations, first for a): Lemma U n, U random variables on (Ω, F, P ) with values in normed vector space (S, ).

8 1.2 Donsker s theorem 8 1. (V n ) n N sequence of S-valued random variables with V n n in probability. If S is separable and U n U, then also U n + V n U. 2. (c n ) n N sequence in R with c n c. If U n U, then also c n U n cu. Proof. 1. S separable B(S S) = B(S) B(S) each U n + V n is still a random variable. Take h C b (S), then 2. E[h(U n + V n ) E[h(U) E[ h(u n + V n ) h(u n ) + E[h(U n ) h(u) }{{} n and E[ h(u n + V n ) h(u n ) 2 h P [ V n δ + sup x,z x z δ h(x) h(z). Now 2 h P [ V n δ as n for each fixed δ (convergence in probability), and sup... h(x) h(z) becomes arbitrarily small for δ small, if h is uniformly continuous. But may take h uniformly continous w.l.o.g., cf. Portmanteau theorem, WT Theorem V.2.2. E[h(c n U n ) E[h(cU) E[ h(c n U n ) h(cu n ) + E[h(cU n ) h(cu) }{{} sup h(x) h(z) + 2 h P [ U n M. x,z x z M c c n Again sup... h(x) h(z) becomes arbitrarily small even for M large, if h is uniformly continuous: ok. For the second term lim sup P [ U n M P [ U M n because U n U, and P [ U M becomes small for M large. Proposition Under the assumptions of Theorem 1.2.1, all fdmds of (µ n ) converge weakly to corresponding fdmds of Wiener measure µ. Proof. Note that even if we do not know yet whether µ exists, we do know how its fdmds must look: they are specified by (BM1) and (BM2 ) as Gaussian. So fix J [, 1 finite and consider µ (J) n = P (Xt n 1,...,t n ) 1, where J = {t 1,..., t m }. First suppose J = {t} (m = 1). Then µ n (J) is the distribution of Xt n, and corresponding fdmd of µ is N (, t). So consider On the one hand X n t = 1 n S [nt + 1 n Y [nt+1 (nt [nt). 1 [nt S [nt S [nt = N (, t), n }{{ n } [nt }{{} t N (,1)

9 1.2 Donsker s theorem 9 by CLT, and on the other hand the second term converges to in probability: 1 Y [nt+1 (nt [nt) n 1 Y [nt+1, n and by Chebyshev P [ Yk δ Var[Y k n nδ 2 n, δ >. So for m = 1, ok. For m > 1 the analogous result follows using CLT in R m. Lemma (Ottaviani). Suppose we have U 1,..., U n independent random variables with E[U i =, n i=1 Var[U i = 1. For Z k = k i=1 U i, we have: [ P for every α > 1. max Z k > 2α k=1,...,n 1 1 1/(α 2 ) P [ Z n > α Proposition Under the assumptions of Theorem 1.2.1, the sequence (µ n ) n N is tight. Proof of Prop We want to use WT, Prop. V.4.3. By construction: n N X n = n N µ n ({X() = }) = 1 µ n = µ n π 1 {} is tight in M 1 (R) It remains to show the second statement in Prop. V.4.3., that is by using: W δ (X) = sup{ X(t) X(s), s, t [, 1, s t δ} and the definition of µ n, that: ε lim lim sup δ n [ P sup s,t s t δ For fixed s with kδ s < (k + 1)δ, we have Xt n Xs n > ε = X n t X n s = X n t X n (k+1)δ + Xn (k+1)δ Xn kδ + X n kδ X n s and t (k + 1)δ δ for t s δ. So we get that: sup Xt n Xs n 3 sup s,t: s t δ k 1 δ sup Xt n Xkδ n t [kδ,(k+1)δ) This gives: P [ sup Xt n Xs n > ε P [ sup sup Xt n Xkδ n > ε/3 s,t s t δ k 1 t [kδ,(k+1)δ) δ P [ sup Xt n Xkδ n > ε/3 k 1/δ t [kδ,(k+1)δ)

10 1.2 Donsker s theorem 1 Fix δ and find, for given n, integers j, m, such that j + m 1 m j m kδ < j + 1 m and (k + 1)δ < j + m m For t that lies between kδ and (k + 1)δ and each lie in turn between two points on a grid of width 1/n. For n sufficiently large we can ensure m/n 2δ, hence m [2δn sup t [kδ,(k+1)δ) { maxl=1,...,m 1 n S j+l S j max l=1,...,m 1 n S j+1+l S j+1 Y k i.i.d. with Var[Y k = 1, S l, l [2δn Var[S l [2δn. Therefore using Lemma with α = ε/6 1/ 2δ: [ [ P Xt n Xkδ n > ε/3 2P max s l > ε/3 n sup t [kδ,(k+1)δ) L (w.l.o.g. δ δ const(ε) independent of δ). By CLT, because S [2δn / [2δn n N (, 1). l=1,...,[2δn [ δ/(ε 2 ) 2P S [2δn [2δn ε 6 2δ [ S[2δn = const(ε)2p [2δn ε 6 2δ ( ( n ε ) ) const(ε)2 1 φ 6 2δ Because we sum over all k 1/δ, we obtain with k(δ) = ε/(6 2δ): lim sup n P [ sup s,t: s t δ For δ, we have k(δ) and because X n t X n s > ε 1 δ const(ε) ε k(δ) k 2 (δ) const(ε) y 2 f(x)dy = 1 6 2δ f(y)dy k(δ) f(x)dy const(ε) y 2 f(x)dy k(δ) the above integral goes to, and this concludes the proof. By combining Proposition and Proposition with WT Thm V.4.3, we n obtain µ n µ and so Donsker s theorem (1.2.1) is proved.

11 1.3 Some applications Some applications Consider situation of Donsker s Theorem: X n, µ n, µ Wiener measure, X BM (so µ n = P (X n ) 1 µ = P X 1 ). Consider F : C[, 1 S (for metric space (S, d)); then F (X n ) F (X) if F is continuous. Lemma D F := {x C[, 1 F is not continuous at x} is in B(C[, 1). Proof. (Note: the proof holds for any metric space, not just C[, 1). F discontinuous at x: ε > such that y, z arbitrarily close to x with d(f (y), F (z)) ε. So: D F = (U δ (y) U δ (z)) B(C[, 1). ε> y,z δ> ε Q δ Q d(f (y),f (z)) ε } {{ } open Proposition Let F : C[, 1 S measurable with µ(d F ) =, i.e., F is µ-a.e. continuous on C[, 1. Under the assumptions of Theorem 1.2.1, we then have F (X n ) F (X) or, equivalently, µ n F 1 µ F 1. Proof. F continuous: obvious; in general, for A S closed, F 1 (A) F 1 (A) D F, so using that F 1 (A) is closed in C[, 1 and the Portmanteau Theorem (WT V.2.2), lim sup n µ n (F 1 (A)) lim sup µ n (F 1 (A)) µ(f 1 (A)) µ(f 1 (A)), n i.e. µ n F 1 µ F 1. How can we use Proposition 1.3.2? Two ways: 1. Suppose we know the distribution of F (X). Then: universal limit result: for any i.i.d. sequence (Y i ) as in Theorem 1.2.1, F (X n ) some limit. 2. To find the distribution of F (X): choose good i.i.d. sequence (Y i ), compute distribution of F (X n ) for that, then compute limit; gives distribution of F (X). Often a good choice is Y i = ±1 with probability 1 2 each. Example.: F (x) = x(1) (final value): then F (X) = X 1 N (, 1) since X is BM. Moreover, F (X n ) = X1 n = 1 S n = 1 n Y i. n n i=1 So Proposition (F is continuous on C[, 1) gives 1 n n i=1 Y i N (, 1),

12 1.3 Some applications 12 whenever (Y i ) are i.i.d. with E[Y i = and Var[Y i = 1. This is the classical CLT. Example.: F (x) := sup t 1 x(t) = max t 1 x(t) (maximum of x): this F is continuous C[, 1 R, indeed even F (x) F (y) x y. Here, F (X) = max t 1 X t is the maximum of BM. Moreover F (X n 1 ) = max t 1 Xn t = max S l l=,...,n n since max will be taken at one of the n + 1 interpolation points. Set M n := max S l = l=,...,n max l=,...,n i= Choose Y i binary: P [Y i = +1 = P [Y i = 1 = 1 2. Lemma For the above binary i.i.d. Y i, l Y i. P [M n b = 2P [S n > b + P [S n = b. Proof. b = : LHS = 1 and RHS = 1 because the distribution of S n is symmetric around. b > : use {S n b} {M n b} to write P [M n b P [S n = b = P [M n b, S n > b + P [M n b, S n < b = P [S n b + P [M n b, S n < b. By reflection principle, P [M n b, S n < b = P [M n b, S n > b, so P [M n b P [S n = b = P [S n b + P [M n b, S n > b = P [S n b + P [S n > b. Example (cont.).: F (X n ) = 1 n M n. So for x and b n := [ nx, Lemma for binary Y i gives [ P [F (X n Sn ) x = P [M n b n = 2P > b [ n Sn + P = b n. n n n n Now take limit as n and use CLT: Sn n N (, 1). Moreover, that P [S n = b n n. Therefore, since X 1 N (, 1). P [ max X t x = P [F (X) x = lim P [F t 1 n (Xn ) x [ = 2 lim P Sn > b n n n n = 2P [X 1 > x = P [ X 1 x, b n n x so Corollary For BM X, max t 1 X t and X 1 have the same distribution, namely P [ max t 1 X t z = P [ X 1 z = 2Φ(z) 1, z [, ).

13 1.3 Some applications 13 t = F (x) y(t) x(t) Figure 2: Example x at which F is discontinuous ±δ u ε F (X) ε F (x) F (X) + ε 1 Figure 3: F must be continuous at a sign change Corollary For every i.i.d. sequence (Y i ) with E[Y i = and Var[Y i = 1, [ l lim P Y i z = 2Φ(z) 1. n max l=,...,n 1 n Proof. Immediate from Corollary and Proposition i=1 Remark.: We could also prove Corollary directly by first proving a reflection principle for BM. This needs the strong Markov property. Example.: F (x) := sup{t [, 1 x(t) = } 1, i.e. F (x) is the time of the last zero of x( ) in [, 1. This F is not continuous on C[, 1: in Figure 2 y x < δ but F (y) = 1 is not close to t < 1, i.e. to F (x). Lemma The above F is µ-a.e. continuous on C[, 1. In other words, F is continuous on typical trajectories of BM; with probability 1, BM cannot behave like x. Proof. Claim: If F is discontinuous at x, then x looks as in Figure 2; more precisely: F discontinuous at x ε > such that x has the same sign on (F (x), 1 and on (F (x) ε, F (x)). Proof of claim: assume x after F (x), but for every ε >, u ε in (F (x) ε, F (x)) with x(u ε ) <. Then F is continuous at x: choose ε >. Choose δ > with x δ on (F (x) + ε, 1. If then y x < δ/2, necessarily F (y) F (x) + ε, since y δ/2 after F (x) + ε. (Figure 3) Take u ε for ε, so x(u ε ) < and u ε lies in (F (x) ε, F (x)). Assume in addition that y x < 1 2 x(u ε) ; then F (y) F (x) ε, since x passes and y is close to x. So F (y) F (x) ε. So define { } A ± x < > on both (F (x), 1 and on := x C[, 1. (F (x) ε, F (x)) for some ε >

14 14 Then D F A + A, so it is enough to prove µ(a ± ) =. Now A { } max (x(s) x(r)) = x(r). r s 1 r Q [,1 For fixed r [, 1, (X s X r ) r s 1 is a BM on [r, 1 and independent of X r. So ({ }) [ µ max (x(s) x(r)) = x(r) = P max (X s X r ) + X r = r s 1 r s 1 and since max r s 1 (X s X r ) =: U N (, 1 r) by Corollary 1.3.7, and X r =: N (, r), where U, V independent, we get ({ }) µ max (x(s) x(r)) = x(r) = P [U + V = =. r s 1 Example.: F (x) := sup{t [, 1 x(t) = } is continuous on C[, 1 (Wiener-measure-)a.e. for the sequence (X n ) from Donsker s theorem, we get F (X n ) F (X). To find the distribution of F (X), again use for X n simple binary random walk; then F (X n ) is the last zero of simple symmetric random walk, and it is known that this has the discrete arcsine distribution. Passing to the limit (and doing some work) we can get Corollary If X is BM and L = F (X) = sup{t [, 1 X t = } is the last zero of X before 1, then L has the arcsine distribution P [L z = 2 π arcsin( z) for z [, For every sequence of i.i.d. random variables (Y i ) with E[Y i = and Var[Y i = 1, we have for all z [, 1: lim P [max{k {,..., n} S k 1S k < } nz = 2 n π arcsin( z) (Details: Billingsley, Sec. 9) 2 Properties of Brownian Motion 2.1 Transformations and First Properties Throughout this section, W = (W t ) t is BM on (Ω, F, P ). How to get new BMs from W? Proposition Each of the following processes is again BM: 1. X 1 := W (reflected at x-axis)

15 2.1 Transformations and First Properties X 2 t := W t+s W s, t, for any fixed s (restarted at time s) 3. X 3 t := cw t/c 2, t, where c > fixed (rescaled in time and space) 4. Inversion of small and large times: { Xt 4 for t = := tw 1/t for t > 5. X 5 t := W t I {t<t} + (W t + Z (W t W t ))I {t t}, where t > is fixed and Z is independent of W and P [Z = +1 = P [Z = 1 = 1 2. Proof. X 5 see exercises. X 1, X 2, X 3 all are P -a.s. continuous. Next, X 1,..., X 4 all are Gaussian processes, all with E[Xt i =. One easily calculates Cov(Xs, i Xt) i = s t; so X 1, X 2, X 3 are BM by Proposition It only remains to prove that X 4 is continuous at, i.e. P [lim t Xt 4 = = 1. Denote by Q the distribution of X 4 on C(, 1. Then Q = µ C(,1, because both X 4 and BM are continuous on (, 1 and, being Gaussian, have the same fdmds. So P [lim t X 4 t = = Q[{x C(, 1 lim t x(t) = } = µ({x C(, 1 lim t x(t) = }) = P [lim t W t = = 1. Corollary (Strong law of large numbers for BM). (for BM W ). W t lim = P -a.s. t t Proof. Wt := tw 1/t is BM by Proposition So [ P lim t W t t [ [ W 1/h = = P lim h 1/h = = P lim h W h = = 1 (last step by Proposition 2.1.1). In particular, Corollary says that BM grows more slowly than linearly as t. So how fast exactly does it grow? Theorem (Law of the iterated logarithm (LIL) for BM). For BM W, with probability 1, lim sup t W t 2t log(log t) = +1, lim inf t W t 2t log(log t) = 1. Theorem is proved like Corollary by using X 4 from Proposition and the following result:

16 2.1 Transformations and First Properties 16 W t t h very small t + h Figure 4: LIL property of BM Theorem (LIL for BM). If we set ψ(h) := have, for BM W, for any t, [ W t+h W t P lim sup = +1 ψ(h) P h [ W t+h W t lim inf = 1 h ψ(h) 2h log(log 1 h ), then we = 1, = 1. So close to the right of t, trajectories of BM behave as sketched in Figure 4. The proof of Theorem is postponed for the moment. Let W be Brownian Motion and define the set of zeroes of W (ω) in [, 1. N(ω) := {t [, 1 W t (ω) = }, Theorem For P -almost all ω, the set N(ω) 1. is closed. 2. has Lebesgue measure. 3. is infinite. 4. is perfect, i.e. it contains no isolated points: each point in N(ω) is a cluster point of points in N(ω). In particular, N(ω) is uncountable. Proof. 1. N(ω) is the set of zeroes of the continuous function W (ω), hence closed. 2. W is product-measurable on Ω [, 1 (exercise), so hence λ(n) = P -a.s. [ 1 E[λ(N) = E I {Wt=} dt = 1 P [W t = }{{} = 3. By the LIL (Theorem 2.1.4), BM oscillates infinitely often close to the right of t =. dt

17 2.2 Path properties of BM We cannot prove this completely (yet), but the argument goes like this: for fixed s, by Proposition 2.1.1, W +s W s is again BM. It seems plausible that this generalizes to random times. So fix q and define the first zero of W after q. τ(ω) := inf{t > q W t (ω) = }, Then we claim (at present without proof) that is again BM. W t := W t+τ W τ = W t+τ, t (2.1.6) By the LIL, t = is a cluster point of the zeros of W from the right, P -a.s. In the same way, by (2.1.6), τ is a cluster point from the right of the zeroes of W, or of W, P -a.s. So the probability that the first zero after some rational q is not a cluster point of zeroes from the right is. But the rational points lie dense in [, 1, and so N(ω) P -a.s. contains no isolated points. That N(ω) is then uncountable is a general fact, see Hewitt/Stromberg Thm. II We still need to prove (2.1.6), and this will be a consequence of the so-called strong Markov property of BM; see later. 2.2 Path properties of BM We know that the trajectories of BM are P -a.s. continuous. But still, they are quite wild: Theorem P -almost all paths W (ω) of BM are nowhere differentiable. Proof. We show that the set of trajectories which are differentiable somewhere has outer measure. (It seems unknown whether this set is measurable.) If W (ω) is differentiable at s, then it has bounded difference quotients close to s, so that W t W s C t s for t near s. Therefore W k n W k 1 n C 1 n, for all n large, k n near s, and for three successive k. So: Now W k+j+1 n 3 P j=1 A := {W is differentiable at some s [, 1 } 3 { W k+j+1 W k+j C 1 } n n n C N m N n m k n j=1 =: B. W k+j n { W k+j+1 n N (, 1 n ), so W k+j n n} C 1 = ( [ P Z C ) 3 const. n 3 2. n

18 2.2 Path properties of BM 18 Next, so B m := n m k n j= P [B m lim sup P n k n j=1 3..., n m, k n j= lim sup n const. n 3 2 n =. So finally P [B =. Immediate consequence: P -almost all paths of BM have infinite variation on any time interval, because Lebesgue s differentiation theorem says that any function of finite variation is λ-a.e. differentiable. Intuitive explanation: W t+h W t N (, h), so in particular, W t+h W t is symmetric around with E[(W t+h W t ) 2 = h. So heuristically, we have W t+h W t is of the order ± h. Now consider a partition of [, 1: Π = {t,..., t k } with t < < t k 1, Π := sup{t k t k 1 t k 1, t k Π}. Then t i Π W ti+1 W ti t ti+1 i Π t i, typically, heuristically. So for a sequence (Π n ) with Π n (e.g. equidistant points), we heuristically expect W t n i+1 W t n n i + t n i Πn since 1/ n is not summable. This explains heuristically why we have infinite variation. In the same heuristic way, we expect: ( Wt n W ) 2 i+1 t n i (t n i+1 t n i ) = 1 t n i Π t n i Πn The surprising thing is that this heuristic reasoning gives a correct result: Theorem (Lévy). Let (Π n ) be a sequence of partitions of [, 1 with Π n. Then for any t [, 1: V n t := t n i Π ( ) 2 n W t t i+1 t W t n i t t in L 2 (P ) If the sequence (Π n ) is refining (i.e., Π n Π n+1 n), then even lim n Vt n = t P -a.s. In other words: P -a.a. trajectories of BM have on [, t quadratic variation t, along any refining sequence of partitions with Π n. Remark.: It is important that the sequence (Π n ) is fixed (it must not depend on ω); for each fixed ω, the quadratic variation of W (ω), computed as the sup of sums of squared increments over all partitions, possibly depending on ω, is +.

19 2.3 Martingale properties of BM 19 Proof. P -a.s. convergence is a bit tricky, see Kallenberg, T L 2 -convergence is easy: take t = 1 and set n i W = W t n i+1 t W t n i t, n t i = t n i+1 t tn i t, so: V n 1 = i ( n i W ) 2. Now: the n i W are independent and N (, n t i ). So E[V n 1 = i n t i = 1, Var[V n 1 = i Var[( n i W ) 2 since: Var[( n i W ) 2 = E[( n i W ) 4 (E[( n i W ) 2 2 ) = 3( n t i ) 2 ( n t i ) 2 we obtain: Hence: Var[V n 1 2 i ( n t i ) 2 Π n 2 n t i. i }{{} =1 V n L 2 = Var[V n 1 n So: along any fixed good sequence of partitions of [, t, BM trajectories have a quadratic variation (of t). This is the key property behind Itô s formula and stochastic calculus; see later. It also implies that BM has infinite variation along any such (good) sequence of partitions, because: any continuous function of finite variation has quadratic variation. (Write n i f := f(tn i+1 t) f(tn i t) to get: ( n i f) 2 sup n i f n i f ; i i i the second factor remains bounded (in n) since f has finite variation, and the first is: sup { f(t n i+1 t) f(t n i t) t n i, t n i+1 Π n} and this goes to as Π n because on [, t, f is uniformly continuous.) 2.3 Martingale properties of BM Start with (Ω, F, P ) and BM (W t ) t on it; so we have (BM1), (BM2 ) and (BM3). There is no filtration yet. Definition. Let H t := σ(w s ; s t) the (raw) filtration generated by W, and: H t := ε> H t+ε Difference: H t H t, and H = (H t ) t is right-continuous; this is important for martingale theory in continuous time.

20 2.3 Martingale properties of BM 2 Lemma For all t and h >, the increment W t+h W t is independent of H t and even of H t. Proof. 1. By (BM2 ), BM has independent increments, and this already implies that W t+h W t is independent of Ht ; see Exercise 2-1. (This uses no properties of distribution of increments nor path properties of BM.) 2. To get from H t to H t, we only use that W has (right-)continuous trajectories, as follows: take g t on Ω, bounded H t -measurable and f on R, bounded, continuous. g t is H t+ε-measurable, W t+h W t+ε is independent of H t+ε. Therefore: E[g t f(w t+h W t+ε ) = E[g t E[f(W t+h W t+ε ). Let ε ; then W t+ε W t P -a.s. by RC, so also f(w t+h W t+ε ) f(w t+h W t ) P -a.s.; all is bounded, therefore Lebesgue gives: E[g t f(w t+h W t ) = E[g t E[f(W t+h W t ) for all f bounded, continuous. Use MCT to extend this to all f bounded measurable, and this shows that W t+h W t is independent of H t. Proposition The following are all (F, P )-martingales: 1) (W t ) t 2) (W 2 t t) t 3) (e αwt 1 2 α2t ) t for any α R. Proof. Adapted: clear. Integrable: W t is normal. Martingale property: 1) E[W t W s F s ind N (,...) = E[W t W s = 2) E[Wt 2 F s = E[(W t W s + W s ) 2 F s W mart. = E[(W t W s ) 2 F s + Ws ind = E[(W t W s ) 2 + Ws 2 N (,t s) = t s + Ws 2 3) With M t := e αwt 1 2 α2t, we get: [ Mt E F s = E[e α(wt Ws) 1 2 α2 (t s) F s M s ind = e 1 2 α2 (t s) E[e α(wt Ws) = e 1 2 α2 (t s) e 1 2 α2 (t s) = 1 To exploit Proposition 2.3.2, we need the stopping theorem for martingales in continuous time. Definition (Stopping time wrt. F). Mapping τ : Ω [, with {τ t} F t, t. σ-field of events observable up to time τ is: check: F τ := {A F A {τ t} F t, t }

21 2.3 Martingale properties of BM 21 W t (ω) a τ a (ω) t Figure 5: Entry time of W into (a, ) no a yes t Figure 6: Local maximum? i) F τ is a σ-field ii) F σ F τ for σ τ iii) If τ t for some fixed t, then F τ = F t. Example.: For a > consider the first entry time of W into (a, ), i.e., τ a := inf{t > W t (a, )} = inf{t > W t > a} (Figure 5). Note: Because W (ω) is continuous, we have W τa = a. This τ a is a stopping time wrt. H (or F), but not wrt. H. Proof. Uses only that τ a is the first time that an adapted RC process (like W ) hits an open set (like (a, )): {τ a t} = {τ a < t + ε} ε> (a, ) is open, W RC = ε> r<t+ε r Q {W r > a} }{{} H r } {{ } H t+ε Now suppose we can only observe H t. τ a t? We cannot tell here: t might be a local maximum of W, but it need not be (Figure 6). This illustrates why τ a cannot be an H -stopping time. Lemma Let X = (X t ) t be adapted and RC and τ be an F-stopping time. Define: X τ : Ω R by (X τ )(ω) := X τ(ω) (ω); assume also that X is well defined (e.g. X := or X := lim n X t ). Then X τ is F τ -measurable. Proof. The lemma is intuitive, but needs proving; see Kallenberg, Lemma 6.5.

22 2.3 Martingale properties of BM 22 Theorem (Stopping Theorem). Let M = (M t ) t be an (F, P ) martingale with RC paths and σ, τ F-stopping times with σ τ. If τ is bounded or if M is uniformly integrable, then: E[M τ F σ = M σ Proof. 1) If M is UI, then the martingale convergence theorem (WT, Thm. IV.3.6) implies that M := lim n M n exists P -a.s., that M L 1 (P ), and that M n = E[M F n, n N. Then also M t = E[M F t, t. If τ N for some N <, then E[M N F t = M t, t, and we only need M on [, N. So in both cases we may assume that M t = E[M N F t, t, and τ N, for some N [,. 2) F σ F τ, conditional expectation is projective it is enough to prove E[M N F τ = M τ. This already follows if ρ E[M ρ is constant over all stopping times ρ. Indeed, take A F τ, ρ N, ρ := τi A + NI A c. Then ρ, ρ are stopping times and so Hence E[M N = E[M ρ = E[M ρ = E[M τ I A + E[M N I A c. E[M N I A = E[M τ I A, A F τ, and M τ is F τ -measurable by Lemma 2.3.4; so M τ = E[M N F τ. 3) Idea: to prove ρ E[M ρ is constant, reduce to discrete time and use stopping theorem from there (WT Thm. IV.3.8). If ρ has only countably many values t n, n N, which are increasingly ordered, view (M tn ) n N as discrete-time martingale in the filtration (F tn ) n N and note that ρ is also a stopping time for this filtration. Stopping Theorem (WT IV.3.8) then says E[M ρ = E[M N (and even M ρ = E[M N F ρ ). Now take a general ρ and define ρ m := (k + 1)2 m I {k2 m ρ<(k+1)2 m } + NI {ρ=n}. k= Then ρ m has only countably many increasingly ordered values, and ρ m is an F-stopping time, because {ρ m t} = {k2 m ρ < (k + 1)2 m } F t for t < N. k:(k+1)2 m t By construction, ρ m ρ, so M ρm M ρ P -a.s. by RC. Moreover, M ρm = E[M N F ρm ; hence (M ρm ) m N is UI and therefore [ E[M ρ = E lim M ρ m = lim E[M ρ m = E[M N. m m So we are done. Remark.: Why assume that F is right-continuous? It seems that we did not use that. General result (Kallenberg, Thm. 7.27): In a right-continuous filtration, every martingale has a version with RCLL trajectories. We always choose such a version (!).

23 2.3 Martingale properties of BM 23 Corollary (Maximal inequality). Let M be a (RC) martingale and set Mt := sup s t M s. Then and letting t then gives [ P sup M s > c s P [Mt > c E[M, c >, c E[M, c >. c Proof. τ c := inf{t M t > c} is a stopping time; {Mt on the set {τ c t}, M τc t = M τc c since M is RC. So > c} {τ c t}, and cp [M t > c = E[cI {M t >c} E[M τc ti {τc t} E[M τc t = E[M. Application.: Estimate for BM W : for any t > and α, β >, [ P (W s α ) 2 s β e αβ. (2.3.7) sup s t Proof. M s := exp(αw s 1 2 α2 s), s, is a martingale by Proposition So [ P sup (W s α ) s t 2 s β = P [Mt e αβ = lim P [Mt > e αγ γ β lim γ β e αγ E[M e αβ. Next we consider the ruin problem for BM. Let W be BM and fix a < < b: W t b τ a,b t a Set with obvious notations. τ a,b := inf{t W t [a, b} = τ a τ b Theorem (Ruin problem for BM). We have P [W τa,b = a = b b a = b b + a, (2.3.9) E[τ a,b = ab = a b. (2.3.1) Proof. We want to use Theorem 2.3.5; but because τ a,b is not bounded and the martingales we examine are not UI, we must be careful.

24 2.4 The law of the iterated logarithm For brevity, write τ := τ a,b. (W 2 t t) t is a martingale by Proposition 2.3.2; so use Theorem for it with bounded stopping time τ n to get E[W 2 τ n = E[τ n E[τ by monotone integration. Now let n : W 2 τ n max(a 2, b 2 ), so E[τ max(a 2, b 2 ) <, so τ < P -a.s. Hence Wτ n 2 Wτ 2 gives E[Wτ 2 = E[τ. So as n P -a.s., and now Lebesgue with p := P [W τa,b = a. E[τ a,b = E[W 2 τ a,b = pa 2 + (1 p)b 2 (2.3.11) 2. Also W is a martingale; so Theorem gives E[W τ n =. As n, W τ n W τ P -a.s. since τ < P -a.s.; and W τ n max( a, b). So Lebesgue gives: E[W τ = lim E[W τ n = n hence = E[W τa,b = pa + (1 p)b Solve this for p to get (2.3.9); plug into (2.3.11) to get (2.3.1). Remark.: By the LIL, we get τ a < P -a.s. and τ b < P -a.s. for any a < < b. Moreover, τ a,b τ b as a. Therefore E[τ b mon. int. = lim E[τ a,b = lim a b = + a a and symmetrically E[τ a =. So these are canonical examples of stopping times τ with τ < P -a.s. but E[τ =. 2.4 The law of the iterated logarithm The goal of this subsection is to prove the LIL (Theorem 2.1.4). First we need a useful technical estimate: Lemma For Z N (, 1) and any a > : 1 a + 1 a e 1 2 a2 2πP [Z > a 1 a e 1 2 a2 That is, the tail probability for N (, 1) decays exponentially with quadratic rate. Proof. RHS: (e 1 2 x2 ) = xe 1 2 x2 ; so: 2πP [Z > a = e 1 2 x2 dx a a x a e 1 2 x 2dx = 1 a e 1 2 a2 LHS: ( 1 x e 1 2 x2 ) = (1 + 1 x )e x2 ; so: πP [Z > a = e 1 2 x2 x dx e 1 2 x2 dx = 1 1 a a e 1 2 a 2. 2 a 2 a

25 2.4 The law of the iterated logarithm 25 Theorem (2.1.4, LIL for BM). Let W be BM and define: ψ(h) := 2h log(log 1 h ). For any t : P [lim sup h P [lim inf h W t+h W t ψ(h) W t+h W t ψ(h) = +1 = 1 = 1 = 1 Proof. (W s+t W t ) s is again BM; so wlog t =. W is also BM; so need only to prove lim sup, that is: { 1 P -a.s. lim sup h W h ψ(h) = 1 P -a.s. Proof of : Idea: approximate ψ(h) in a piecewise affine way, and estimate W very well against these affine functions. Choose a sequence of the form h n = θ n 1 with < θ < 1; set α n := (1 + δ) h n+1 ψ(h n+1 ), β n := 1 2 ψ(h n+1) with δ >. Then: P [ sup (W s α n s h n 2 s) β n (2.3.7) = e αnβn =... = const(n + 1) (1+δ) This is summable over n since δ > ; so Borel-Cantelli implies that with probability 1, sup s hn (W s αn 2 s) β n finally, i.e., for n large enough. So we have P -a.s. for n n (ω) that if we choose h > with h n+1 < h h n, then: W h sup (W s α n s h n 2 s) + α n 2 h n β n + α n 2 h n now use h n = 1 θ h n+1 and the definition of α n, β n ; then P -a.s., for n n (ω), for h with h n+1 < h h n : ( 1 W h ψ(h n+1 ) δ ) ( 1 ψ(h) 2θ δ ) 2θ since ψ is increasing for small h. Take the lim sup as h ; then let δ, θ 1 and get: W h lim sup h ψ(h) 1 P -a.s. Proof of : write W hn = W hn W hn+1 +W hn+1 ; estimate independent random variables W hn W hn+1 from below; use 1) for W to also estimate W hn+1 from below; this gives estimate from below for W hn. Again choose h n = θ n with < θ < 1. By 1), applied to W, we have P -a.s. for n n 1 (ω) W hn+1 2ψ(h n+1 ) (2.4.3)

26 26 On the other hand, W hn W hn+1 =: h n h n+1 Y n = θ n θ n+1 Y n, and the Y n are i.i.d. N (, 1). To estimate P [W hn W hn+1 Cψ(h n ), we choose Now a n := C θn (1 θ) ψ(h n) = = C 2 log n + const.(θ). 1 θ 1 P [W hn W hn+1 Cψ(h n ) = P [Y n a n a n + 1 e 1 a n by Lemma Compute for large n, and exp for C := 1 θ. Therefore n=1 a n + 1 a n const. log n const. log n, ( 1 ) 2 a2 n =... = const. n C2 1 1 θ = const. n n=1 2 a2 n 1 2π P [W hn W hn θψ(h n ) const. n log n = +. By Borel-Cantelli: P -a.s., often W hn W hn+1 1 θψ(h n ). Combine this with (2.4.3); then P -a.s., for many n, W hn = W hn W hn+1 + W hn+1 1 θψ(h n ) 2ψ(h n+1 ) ψ(h n )( 1 θ 4 θ), using that ψ(h n+1 ) 2 θψ(h n ) for large n. So lim sup h W h ψ(h) 1 θ 4 θ P -a.s. and now the assertion follows as θ. 3 Markov processes Basic idea: Markov process indexed by time is a model for temporal evolution of a stochastic system, where predictions about the future do not depend on the entire past, but only on the present state of the system. More mathematically, we want E[g(X u ; u t) σ(x s ; s t) = E[g(X u ; u t) σ(x t ). In addition, we want this not only for fixed times t, but also for random (stopping) times τ, a property called the strong Markov property.

27 3.1 Basic concepts Basic concepts Start with probability space (Ω, F, P ) and measurable space (S, S); the latter is the state space of the process we want to consider. Suppose we have an S-valued stochastic process X = (X t ) t and define Ft := σ(x s ; s t), t. If S is Polish and S = B(S), then X t has a regular conditional distribution given Fs for any s t. If this depends not on all of Fs, but only on X s, we can write P [X t A F s = P [X t A X s = K s,t (X s, A) for some stochastic kernel K s,t from (S, S) to itself. Moreover, for s t u, K s,u (X s, A) = P [X u A Fs = E [ P [X u A Ft F s }{{} = S K t,u (y, A)K s,t (X s, dy) =: (K s,t K t,u )(X s, A) K t,u(x t,a) If things are homogeneous in time, then K s,t will not depend separately on s, t, but only on time difference t s. For many issues, one needs to work with Markov processes having nice trajectories; so we need a nice function space. So assume (S, d) is a metric space and denote by D(S) the space of all functions y : [, ) S which are RCLL. Call Y the coordinate process on D(S): Y = (Y t ) t with Y t : D(S) S, y Y t (y) = y(t), and D t := σ(y s ; s t), D := D = σ(y s ; s ) (on D(S)). Further we define S [, ) := σ(y s ; s ) (on S [, ) ). Then (D(S), D) is a measurable space; it will turn out that if S is nice, then also D(S) is nice. In other words: RCLL processes are nice objects. One can view an S-valued stochastic process X = (X t ) t on (Ω, F, P ) as an S [, ) -valued random variable; if X has RCLL trajectories, it is actually D(S)-valued. Distribution of X under P is a p.m. on (S [, ), S [, ) ) or on (D(S), D), depending on trajectories of X. Theorem Suppose S is Polish and S = B(S); then there is a metric on D(S) which makes D(S) into a complete separable metric space (so D(S) is again Polish) and such that D = B(D(S)). [So: completely analogous to C(S) The resulting topology is the Skorohod topology. Proof. Long and technical; see Ethier/Kurtz, section 3.5, or Jacod/Shiryaev, Section VI.1. Definition. (S, d) metric space, S = B(S). A transition semigroup on (S, S) is a family (K t ) t of stochastic kernels on (S, S) such that: (TS1) K = Id, i.e., K (x, A) = I A (x), x S, A S. (TS2) K s K t = K t+s (= K t K s ) for s, t, where (K s K t )(x, A) := S K s(x, dy)k t (y, A). This is called the semigroup property or Chapman-Kolmogorov equation. (TS3) lim h K h f(x) = f(x), x S, f C b (S), where (K h f)(x) := S K h(x, dy)f(y) for f measurable and bounded or.

28 3.1 Basic concepts 28 Note.: View K t interchangeably as stochastic kernel and as operator acting on functions. Then (TS3) says that (K t ) t is continuous at t =, with respect to pointwise convergence for functions in C b (S); also called in bp-sense. Definition. (S, d) metric space, S = B(S). A (RCLL) Markov process on (Ω, F, P ), with respect to the filtration G = (G t ) t on (Ω, F), with transition semigroup (K t ) is a (RCLL) S-valued stochastic process X = (X t ) t which is adapted to G and such that for all t, h, the conditional distribution of X t+h given G t is given by K h (X t, ), i.e., P [X t+h A G t = K h (X t, A) P -a.s. The distribution on S [, ) (or D(S)) of X under P is then called a Markov distribution with transition semigroup (K t ), and the distribution on S of X under P is called initial distribution. Remark.: 1) Rigorous formulation of intuitive idea at beginning: K h describes the transition mechanism; if we begin at x S, then K h (x, A) is conditional probability of landing in A after h units of time, given that we start in x. If we start at time t, this conditional probability does not depend on G t, but only on starting point X t. 2) Homogeneous in time: transition mechanism from t to u depends only on u t, not on t, u separately; otherwise, need (K s,t ) s t, and the semigroup property is then K s,t K t,u = K s,u. 3) In discrete time with (X n ) n N, a Markov process is usually called Markov chain. There (at least for the time-homogeneous case), it is enough to specify one single kernel, namely K 1 (passage from one time point to next one). 4) If X is Markov wrt. G, it is also Markov with respect to its own filtration F (projectivity); in particular, need G F (this is implicit if X G-adapted). Without mention of filtration, default is always F. 5) Being Markov is a distributional property (and has nothing to do with paths). In particular if X is P -Markov it need not be Q-Markov for some other p.m. Q on (Ω, F). Notation.: The distribution of X under P, initial distribution, on S is usually called ν. The distribution of X with initial distribution ν is called P ν ; if ν = δ {x}, briefly P x for P δ{x}. Exercise.: X Markov for F, f measurable on S, then E[f(X t+h ) Ft = f(y)k h (X t, dy) = (K h f)(x t ). S Remark.: Suppose X is Markov. Take any Z on S [, ) (or D(S)), Z, S [, ) - (or D-) measurable. Then S x E x [Z is measurable; see below. Example.: If W is BM wrt. G, then W is a continuous Markov process wrt. G: W t+h W t is independent of G t and N (, h), so W t+h = W t + (W t+h W t ), conditionally on G t, has a N (W t, h)-distribution; so K h (x, dz) = n x,h (z)dz = 1 e (z x)2 2h dz. 2πh Check: this family (K t ) t is a transition semigroup.

29 3.1 Basic concepts 29 Example.: A Homogeneous Poisson process with intensity (rate) λ > is a stochastic process N with RCLL, piecewise constant trajectories, with jumps of height +1 at random times in such a way that the increments of N are independent, with N t N s P(λ(t s)). This is an RCLL Markov process; state space is R; since N t+h = N t +(N t+h N t ), kernels are given by: λh (λh)j K h (x, {x + j}) = e j! for x R, j N. What is the structure of a Markov distribution on path space? Proposition A process X is Markov with transition semigroup (K t ) and initial distribution ν iff: n N, = t < t 1 < < t n <, f, f 1,..., f n : S [, ) measurable, we have [ n E k= f k (X tk ) = S S ν(dx )f (x ) S K t1 (x, dx 1 )f 1 (x 1 ) K t2 t 1 (x 1, dx 2 )f 2 (x 2 ) S K tn t n 1 (x n 1, dx n )f n (x n ). (3.1.5) This is equivalent to saying that the marginal distribution of P ν at time points t, t 1,..., t n is given by ν K t1 K t2 t 1 K tn t n 1 = P (X t, X t1,..., X tn ) 1. (3.1.6) Proof. First, (3.1.5) determines marginal distribution of (X t, X t1,..., X tn ); so knowing all (3.1.5) is equivalent to knowing P ν. Second, (3.1.5) and (3.1.6) are equivalent by construction of P 1 K; see WT, Chapter II. Suppose X is Markov. Then So E[f n (X tn ) F t n 1 = (K tn t n 1 f n )(X tn 1 ) =: f n 1 (X tn 1 ). [ n E k= [n 1 f k (X tk ) = E k= f k (X tk )E[f n (X tn ) Ft n 1 n 1 = E[ fn 1 (X tn 1 ) f k (X tk ). Now iterate and use X t = X ν to get (3.1.5). k= Conversely, suppose we have (3.1.5). Then we need to show that E[f(X t+h )Z = E[((K h f)(x t ))Z, Z F t -measurable. Use MCT with M := {all products m i= f i(x ti )} where t m t and f i : S [, ) bounded measurable enough to prove [ E f(x t+h ) m i= [ f i (X ti ) = E (K h f)(x t ) m i= f i (X ti ) ; and this is immediate from the definition of K h f and from (3.1.5).

30 3.2 Markov property and strong Markov property 3 Next question: How can we actually construct a Markov distribution? Obvious idea: define fdmds by (3.1.5) or (3.1.6) and check consistency. This works! Intuition: (3.1.5) is roughly P [ X t (z, z + dz,..., X tn (z n, z n + dz n = ν(dz )K t1 (z, dz 1 )K t2 t 1 (z 1, dz 2 ) K tn t n 1 (z n 1, dz n ) i.e. choose a starting point z according to ν; then successively pass from z k 1 at t k 1 to some z k at t k according to K tk t k 1 (z k 1, dz k ). Now how about existence of Markov processes? Proposition Suppose S is Polish and S = B(S). Take any initial distribution ν and any transition semigroup (K t ) t on (S, S). Then Markov process correpsonding to ν, (K t ); more precisely, choose Ω := S [, ), X := Y := coordinate process, Ft := σ(x s ; s t), F := F := σ(x s ; s ); then,! probability measure P on (Ω, F) such that X is (P, F )-Markov with initial distribution ν and transition semigroup (K t ). Proof. For each finite I [, ), define probability measure Q (I) on (S I, S I ) by RHS of (3.1.6). Then (TS2) (the semigroup property K t K s = K t+s ) implies that the Q (I) are consistent (exercise). So the result follows from Kolmogorov s consistency theorem, WT II.3.5, and of course from Proposition Remark.: 1. For ν = δ {x}, then (3.1.6) implies that x E x [Z is measurable for any measurable Z on S [, ) ; see WT, construction of P 1 K. 2. Without extra conditions on (K t ), X from Proposition will not have nice paths; in other words, P lives on all of S [, ), it is not concentrated on D(S). If (K t ) is Feller (see later), things look better. 3.2 Markov property and strong Markov property Intuition for MP: predictions about future given past only depend on present. This is a distributional property, and is most conveniently formulated on path space. So: Ω = S [, ), X = Y = coordinate process; F = filtration generated by X; F = F ; P = P ν for some initial distribution ν and transition semigroup (K t ). (If we need RCLL paths for X, then work on D(S) instead; assume then that (S, d) is metric and S = B(S).) Definition. Shift operator θ t : Ω Ω (for t ) is defined by (θ t ω)(s) := ω(t + s), i.e. θ t ω( ) = ω(t + ): look at ω only on [t, ). For a mapping τ : Ω [,, define θ τ on {τ < } by θ τ ω := θ τ(ω) ω. Past of X up to time t is F t = σ(x s ; s t). Future of X after time t is ˆF t := σ(x u ; u t).

31 3.2 Markov property and strong Markov property 31 Then θ 1 t (F ) = θt 1 (σ(x s ; s )) = σ(θt 1 (X s ); s ) = σ(x u ; u t) = }{{} ˆF X t+s so θ t is ˆF t -F -measurable, and more precisely all ˆF t -measurable random variables are of the form Z θ t for some F -measurable Z, and vice versa. Proposition (Markov property). For any t and any F -measurable Z on Ω, we have E ν [Z θ t F t = E Xt [Z := E x [Z x=xt P ν -a.s. (3.2.2) Remark.: Z θ t on LHS is a typical ˆF t -measurable random variable; so LHS is the typical prediction of the future given the past. x E x [Z is measurable; so RHS is a measurable function of current state X t and so σ(x t )-measurable. Conditioning on σ(x t ) therefore gives E ν [Z θ t F t = E ν [Z θ t σ(x t ), and the RHS only depends on present state X t, or on present σ(x t ). In particular, Z := I Xh A gives Z θ t = I {Xt+h A} and: P ν [X t+h A F t = E ν [Z θ t F t (3.2.2) = E Xt [Z = P Xt [X h A (3.1.6) = K h (X t, A) So (3.2.2) yields interpretation of K h as conditional distribution of X t+h given F t. The next proof shows that, conversely, (3.2.2) is just a reformulation of this. Proof of Need to show, for every U F t -measurable, E ν [(Z θ t )U = E ν [E Xt [ZU By MCT (..1), it is enough to take U of the form U = n i= f i(x ti ) with f i measurable and all t i t, and for Z of the form Z = m k= g k(x tk ) with g k measurable and all t k. And then this equality follows directly from (3.1.5) in Proposition Example (Short-rate models in mathematical finance).: Suppose r = (r t ) t describes instantaneous interest rate. A zero coupon bond with maturity T pays out at time T > amount 1; price of this at time t < T is given by: [ B t,t = E exp ( T r s ds ) G t t for some Q P, if we want to avoid arbitrage. t ; (3.2.4)

Brownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539

Brownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539 Brownian motion Samy Tindel Purdue University Probability Theory 2 - MA 539 Mostly taken from Brownian Motion and Stochastic Calculus by I. Karatzas and S. Shreve Samy T. Brownian motion Probability Theory

More information

1. Stochastic Processes and filtrations

1. Stochastic Processes and filtrations 1. Stochastic Processes and 1. Stoch. pr., A stochastic process (X t ) t T is a collection of random variables on (Ω, F) with values in a measurable space (S, S), i.e., for all t, In our case X t : Ω S

More information

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS PROBABILITY: LIMIT THEOREMS II, SPRING 218. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please

More information

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3 Brownian Motion Contents 1 Definition 2 1.1 Brownian Motion................................. 2 1.2 Wiener measure.................................. 3 2 Construction 4 2.1 Gaussian process.................................

More information

Lecture 17 Brownian motion as a Markov process

Lecture 17 Brownian motion as a Markov process Lecture 17: Brownian motion as a Markov process 1 of 14 Course: Theory of Probability II Term: Spring 2015 Instructor: Gordan Zitkovic Lecture 17 Brownian motion as a Markov process Brownian motion is

More information

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS PROBABILITY: LIMIT THEOREMS II, SPRING 15. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please

More information

Lecture 21 Representations of Martingales

Lecture 21 Representations of Martingales Lecture 21: Representations of Martingales 1 of 11 Course: Theory of Probability II Term: Spring 215 Instructor: Gordan Zitkovic Lecture 21 Representations of Martingales Right-continuous inverses Let

More information

Lecture 12. F o s, (1.1) F t := s>t

Lecture 12. F o s, (1.1) F t := s>t Lecture 12 1 Brownian motion: the Markov property Let C := C(0, ), R) be the space of continuous functions mapping from 0, ) to R, in which a Brownian motion (B t ) t 0 almost surely takes its value. Let

More information

Functional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals

Functional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals Functional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals Noèlia Viles Cuadros BCAM- Basque Center of Applied Mathematics with Prof. Enrico

More information

MATH 6605: SUMMARY LECTURE NOTES

MATH 6605: SUMMARY LECTURE NOTES MATH 6605: SUMMARY LECTURE NOTES These notes summarize the lectures on weak convergence of stochastic processes. If you see any typos, please let me know. 1. Construction of Stochastic rocesses A stochastic

More information

µ X (A) = P ( X 1 (A) )

µ X (A) = P ( X 1 (A) ) 1 STOCHASTIC PROCESSES This appendix provides a very basic introduction to the language of probability theory and stochastic processes. We assume the reader is familiar with the general measure and integration

More information

4th Preparation Sheet - Solutions

4th Preparation Sheet - Solutions Prof. Dr. Rainer Dahlhaus Probability Theory Summer term 017 4th Preparation Sheet - Solutions Remark: Throughout the exercise sheet we use the two equivalent definitions of separability of a metric space

More information

Wiener Measure and Brownian Motion

Wiener Measure and Brownian Motion Chapter 16 Wiener Measure and Brownian Motion Diffusion of particles is a product of their apparently random motion. The density u(t, x) of diffusing particles satisfies the diffusion equation (16.1) u

More information

Lecture 9. d N(0, 1). Now we fix n and think of a SRW on [0,1]. We take the k th step at time k n. and our increments are ± 1

Lecture 9. d N(0, 1). Now we fix n and think of a SRW on [0,1]. We take the k th step at time k n. and our increments are ± 1 Random Walks and Brownian Motion Tel Aviv University Spring 011 Lecture date: May 0, 011 Lecture 9 Instructor: Ron Peled Scribe: Jonathan Hermon In today s lecture we present the Brownian motion (BM).

More information

I forgot to mention last time: in the Ito formula for two standard processes, putting

I forgot to mention last time: in the Ito formula for two standard processes, putting I forgot to mention last time: in the Ito formula for two standard processes, putting dx t = a t dt + b t db t dy t = α t dt + β t db t, and taking f(x, y = xy, one has f x = y, f y = x, and f xx = f yy

More information

1 Independent increments

1 Independent increments Tel Aviv University, 2008 Brownian motion 1 1 Independent increments 1a Three convolution semigroups........... 1 1b Independent increments.............. 2 1c Continuous time................... 3 1d Bad

More information

Filtrations, Markov Processes and Martingales. Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition

Filtrations, Markov Processes and Martingales. Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition Filtrations, Markov Processes and Martingales Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition David pplebaum Probability and Statistics Department,

More information

Notes 1 : Measure-theoretic foundations I

Notes 1 : Measure-theoretic foundations I Notes 1 : Measure-theoretic foundations I Math 733-734: Theory of Probability Lecturer: Sebastien Roch References: [Wil91, Section 1.0-1.8, 2.1-2.3, 3.1-3.11], [Fel68, Sections 7.2, 8.1, 9.6], [Dur10,

More information

Brownian Motion and Conditional Probability

Brownian Motion and Conditional Probability Math 561: Theory of Probability (Spring 2018) Week 10 Brownian Motion and Conditional Probability 10.1 Standard Brownian Motion (SBM) Brownian motion is a stochastic process with both practical and theoretical

More information

1. Aufgabenblatt zur Vorlesung Probability Theory

1. Aufgabenblatt zur Vorlesung Probability Theory 24.10.17 1. Aufgabenblatt zur Vorlesung By (Ω, A, P ) we always enote the unerlying probability space, unless state otherwise. 1. Let r > 0, an efine f(x) = 1 [0, [ (x) exp( r x), x R. a) Show that p f

More information

ELEMENTS OF PROBABILITY THEORY

ELEMENTS OF PROBABILITY THEORY ELEMENTS OF PROBABILITY THEORY Elements of Probability Theory A collection of subsets of a set Ω is called a σ algebra if it contains Ω and is closed under the operations of taking complements and countable

More information

Stochastic Processes II/ Wahrscheinlichkeitstheorie III. Lecture Notes

Stochastic Processes II/ Wahrscheinlichkeitstheorie III. Lecture Notes BMS Basic Course Stochastic Processes II/ Wahrscheinlichkeitstheorie III Michael Scheutzow Lecture Notes Technische Universität Berlin Sommersemester 218 preliminary version October 12th 218 Contents

More information

Stochastic Processes. Winter Term Paolo Di Tella Technische Universität Dresden Institut für Stochastik

Stochastic Processes. Winter Term Paolo Di Tella Technische Universität Dresden Institut für Stochastik Stochastic Processes Winter Term 2016-2017 Paolo Di Tella Technische Universität Dresden Institut für Stochastik Contents 1 Preliminaries 5 1.1 Uniform integrability.............................. 5 1.2

More information

Brownian Motion. Chapter Definition of Brownian motion

Brownian Motion. Chapter Definition of Brownian motion Chapter 5 Brownian Motion Brownian motion originated as a model proposed by Robert Brown in 1828 for the phenomenon of continual swarming motion of pollen grains suspended in water. In 1900, Bachelier

More information

{σ x >t}p x. (σ x >t)=e at.

{σ x >t}p x. (σ x >t)=e at. 3.11. EXERCISES 121 3.11 Exercises Exercise 3.1 Consider the Ornstein Uhlenbeck process in example 3.1.7(B). Show that the defined process is a Markov process which converges in distribution to an N(0,σ

More information

Preliminary Exam: Probability 9:00am 2:00pm, Friday, January 6, 2012

Preliminary Exam: Probability 9:00am 2:00pm, Friday, January 6, 2012 Preliminary Exam: Probability 9:00am 2:00pm, Friday, January 6, 202 The exam lasts from 9:00am until 2:00pm, with a walking break every hour. Your goal on this exam should be to demonstrate mastery of

More information

Applications of Ito s Formula

Applications of Ito s Formula CHAPTER 4 Applications of Ito s Formula In this chapter, we discuss several basic theorems in stochastic analysis. Their proofs are good examples of applications of Itô s formula. 1. Lévy s martingale

More information

Large Sample Theory. Consider a sequence of random variables Z 1, Z 2,..., Z n. Convergence in probability: Z n

Large Sample Theory. Consider a sequence of random variables Z 1, Z 2,..., Z n. Convergence in probability: Z n Large Sample Theory In statistics, we are interested in the properties of particular random variables (or estimators ), which are functions of our data. In ymptotic analysis, we focus on describing the

More information

(B(t i+1 ) B(t i )) 2

(B(t i+1 ) B(t i )) 2 ltcc5.tex Week 5 29 October 213 Ch. V. ITÔ (STOCHASTIC) CALCULUS. WEAK CONVERGENCE. 1. Quadratic Variation. A partition π n of [, t] is a finite set of points t ni such that = t n < t n1

More information

n E(X t T n = lim X s Tn = X s

n E(X t T n = lim X s Tn = X s Stochastic Calculus Example sheet - Lent 15 Michael Tehranchi Problem 1. Let X be a local martingale. Prove that X is a uniformly integrable martingale if and only X is of class D. Solution 1. If If direction:

More information

Lecture 19 L 2 -Stochastic integration

Lecture 19 L 2 -Stochastic integration Lecture 19: L 2 -Stochastic integration 1 of 12 Course: Theory of Probability II Term: Spring 215 Instructor: Gordan Zitkovic Lecture 19 L 2 -Stochastic integration The stochastic integral for processes

More information

Verona Course April Lecture 1. Review of probability

Verona Course April Lecture 1. Review of probability Verona Course April 215. Lecture 1. Review of probability Viorel Barbu Al.I. Cuza University of Iaşi and the Romanian Academy A probability space is a triple (Ω, F, P) where Ω is an abstract set, F is

More information

An Introduction to Stochastic Processes in Continuous Time

An Introduction to Stochastic Processes in Continuous Time An Introduction to Stochastic Processes in Continuous Time Flora Spieksma adaptation of the text by Harry van Zanten to be used at your own expense May 22, 212 Contents 1 Stochastic Processes 1 1.1 Introduction......................................

More information

Exercises in stochastic analysis

Exercises in stochastic analysis Exercises in stochastic analysis Franco Flandoli, Mario Maurelli, Dario Trevisan The exercises with a P are those which have been done totally or partially) in the previous lectures; the exercises with

More information

Poisson random measure: motivation

Poisson random measure: motivation : motivation The Lévy measure provides the expected number of jumps by time unit, i.e. in a time interval of the form: [t, t + 1], and of a certain size Example: ν([1, )) is the expected number of jumps

More information

Random Process Lecture 1. Fundamentals of Probability

Random Process Lecture 1. Fundamentals of Probability Random Process Lecture 1. Fundamentals of Probability Husheng Li Min Kao Department of Electrical Engineering and Computer Science University of Tennessee, Knoxville Spring, 2016 1/43 Outline 2/43 1 Syllabus

More information

9 Brownian Motion: Construction

9 Brownian Motion: Construction 9 Brownian Motion: Construction 9.1 Definition and Heuristics The central limit theorem states that the standard Gaussian distribution arises as the weak limit of the rescaled partial sums S n / p n of

More information

Convergence of Feller Processes

Convergence of Feller Processes Chapter 15 Convergence of Feller Processes This chapter looks at the convergence of sequences of Feller processes to a iting process. Section 15.1 lays some ground work concerning weak convergence of processes

More information

X n D X lim n F n (x) = F (x) for all x C F. lim n F n(u) = F (u) for all u C F. (2)

X n D X lim n F n (x) = F (x) for all x C F. lim n F n(u) = F (u) for all u C F. (2) 14:17 11/16/2 TOPIC. Convergence in distribution and related notions. This section studies the notion of the so-called convergence in distribution of real random variables. This is the kind of convergence

More information

The strictly 1/2-stable example

The strictly 1/2-stable example The strictly 1/2-stable example 1 Direct approach: building a Lévy pure jump process on R Bert Fristedt provided key mathematical facts for this example. A pure jump Lévy process X is a Lévy process such

More information

Empirical Processes: General Weak Convergence Theory

Empirical Processes: General Weak Convergence Theory Empirical Processes: General Weak Convergence Theory Moulinath Banerjee May 18, 2010 1 Extended Weak Convergence The lack of measurability of the empirical process with respect to the sigma-field generated

More information

Exercises. T 2T. e ita φ(t)dt.

Exercises. T 2T. e ita φ(t)dt. Exercises. Set #. Construct an example of a sequence of probability measures P n on R which converge weakly to a probability measure P but so that the first moments m,n = xdp n do not converge to m = xdp.

More information

In terms of measures: Exercise 1. Existence of a Gaussian process: Theorem 2. Remark 3.

In terms of measures: Exercise 1. Existence of a Gaussian process: Theorem 2. Remark 3. 1. GAUSSIAN PROCESSES A Gaussian process on a set T is a collection of random variables X =(X t ) t T on a common probability space such that for any n 1 and any t 1,...,t n T, the vector (X(t 1 ),...,X(t

More information

6. Brownian Motion. Q(A) = P [ ω : x(, ω) A )

6. Brownian Motion. Q(A) = P [ ω : x(, ω) A ) 6. Brownian Motion. stochastic process can be thought of in one of many equivalent ways. We can begin with an underlying probability space (Ω, Σ, P) and a real valued stochastic process can be defined

More information

STAT 7032 Probability Spring Wlodek Bryc

STAT 7032 Probability Spring Wlodek Bryc STAT 7032 Probability Spring 2018 Wlodek Bryc Created: Friday, Jan 2, 2014 Revised for Spring 2018 Printed: January 9, 2018 File: Grad-Prob-2018.TEX Department of Mathematical Sciences, University of Cincinnati,

More information

Strong approximation for additive functionals of geometrically ergodic Markov chains

Strong approximation for additive functionals of geometrically ergodic Markov chains Strong approximation for additive functionals of geometrically ergodic Markov chains Florence Merlevède Joint work with E. Rio Université Paris-Est-Marne-La-Vallée (UPEM) Cincinnati Symposium on Probability

More information

Lecture 5. If we interpret the index n 0 as time, then a Markov chain simply requires that the future depends only on the present and not on the past.

Lecture 5. If we interpret the index n 0 as time, then a Markov chain simply requires that the future depends only on the present and not on the past. 1 Markov chain: definition Lecture 5 Definition 1.1 Markov chain] A sequence of random variables (X n ) n 0 taking values in a measurable state space (S, S) is called a (discrete time) Markov chain, if

More information

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R. Ergodic Theorems Samy Tindel Purdue University Probability Theory 2 - MA 539 Taken from Probability: Theory and examples by R. Durrett Samy T. Ergodic theorems Probability Theory 1 / 92 Outline 1 Definitions

More information

Weak convergence and Brownian Motion. (telegram style notes) P.J.C. Spreij

Weak convergence and Brownian Motion. (telegram style notes) P.J.C. Spreij Weak convergence and Brownian Motion (telegram style notes) P.J.C. Spreij this version: December 8, 2006 1 The space C[0, ) In this section we summarize some facts concerning the space C[0, ) of real

More information

LECTURE 2: LOCAL TIME FOR BROWNIAN MOTION

LECTURE 2: LOCAL TIME FOR BROWNIAN MOTION LECTURE 2: LOCAL TIME FOR BROWNIAN MOTION We will define local time for one-dimensional Brownian motion, and deduce some of its properties. We will then use the generalized Ray-Knight theorem proved in

More information

PROBABILITY THEORY II

PROBABILITY THEORY II Ruprecht-Karls-Universität Heidelberg Institut für Angewandte Mathematik Prof. Dr. Jan JOHANNES Outline of the lecture course PROBABILITY THEORY II Summer semester 2016 Preliminary version: April 21, 2016

More information

Fundamental Inequalities, Convergence and the Optional Stopping Theorem for Continuous-Time Martingales

Fundamental Inequalities, Convergence and the Optional Stopping Theorem for Continuous-Time Martingales Fundamental Inequalities, Convergence and the Optional Stopping Theorem for Continuous-Time Martingales Prakash Balachandran Department of Mathematics Duke University April 2, 2008 1 Review of Discrete-Time

More information

4 Sums of Independent Random Variables

4 Sums of Independent Random Variables 4 Sums of Independent Random Variables Standing Assumptions: Assume throughout this section that (,F,P) is a fixed probability space and that X 1, X 2, X 3,... are independent real-valued random variables

More information

Part III Advanced Probability

Part III Advanced Probability Part III Advanced Probability Based on lectures by M. Lis Notes taken by Dexter Chua Michaelmas 2017 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after

More information

GAUSSIAN PROCESSES; KOLMOGOROV-CHENTSOV THEOREM

GAUSSIAN PROCESSES; KOLMOGOROV-CHENTSOV THEOREM GAUSSIAN PROCESSES; KOLMOGOROV-CHENTSOV THEOREM STEVEN P. LALLEY 1. GAUSSIAN PROCESSES: DEFINITIONS AND EXAMPLES Definition 1.1. A standard (one-dimensional) Wiener process (also called Brownian motion)

More information

MAT 570 REAL ANALYSIS LECTURE NOTES. Contents. 1. Sets Functions Countability Axiom of choice Equivalence relations 9

MAT 570 REAL ANALYSIS LECTURE NOTES. Contents. 1. Sets Functions Countability Axiom of choice Equivalence relations 9 MAT 570 REAL ANALYSIS LECTURE NOTES PROFESSOR: JOHN QUIGG SEMESTER: FALL 204 Contents. Sets 2 2. Functions 5 3. Countability 7 4. Axiom of choice 8 5. Equivalence relations 9 6. Real numbers 9 7. Extended

More information

MARTINGALES: VARIANCE. EY = 0 Var (Y ) = σ 2

MARTINGALES: VARIANCE. EY = 0 Var (Y ) = σ 2 MARTINGALES: VARIANCE SIMPLEST CASE: X t = Y 1 +... + Y t IID SUM V t = X 2 t σ 2 t IS A MG: EY = 0 Var (Y ) = σ 2 V t+1 = (X t + Y t+1 ) 2 σ 2 (t + 1) = V t + 2X t Y t+1 + Y 2 t+1 σ 2 E(V t+1 F t ) =

More information

Lecture 2. We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales.

Lecture 2. We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales. Lecture 2 1 Martingales We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales. 1.1 Doob s inequality We have the following maximal

More information

Solution for Problem 7.1. We argue by contradiction. If the limit were not infinite, then since τ M (ω) is nondecreasing we would have

Solution for Problem 7.1. We argue by contradiction. If the limit were not infinite, then since τ M (ω) is nondecreasing we would have 362 Problem Hints and Solutions sup g n (ω, t) g(ω, t) sup g(ω, s) g(ω, t) µ n (ω). t T s,t: s t 1/n By the uniform continuity of t g(ω, t) on [, T], one has for each ω that µ n (ω) as n. Two applications

More information

Introduction to Random Diffusions

Introduction to Random Diffusions Introduction to Random Diffusions The main reason to study random diffusions is that this class of processes combines two key features of modern probability theory. On the one hand they are semi-martingales

More information

A D VA N C E D P R O B A B I L - I T Y

A D VA N C E D P R O B A B I L - I T Y A N D R E W T U L L O C H A D VA N C E D P R O B A B I L - I T Y T R I N I T Y C O L L E G E T H E U N I V E R S I T Y O F C A M B R I D G E Contents 1 Conditional Expectation 5 1.1 Discrete Case 6 1.2

More information

Math 6810 (Probability) Fall Lecture notes

Math 6810 (Probability) Fall Lecture notes Math 6810 (Probability) Fall 2012 Lecture notes Pieter Allaart University of North Texas September 23, 2012 2 Text: Introduction to Stochastic Calculus with Applications, by Fima C. Klebaner (3rd edition),

More information

Some Background Material

Some Background Material Chapter 1 Some Background Material In the first chapter, we present a quick review of elementary - but important - material as a way of dipping our toes in the water. This chapter also introduces important

More information

STAT 331. Martingale Central Limit Theorem and Related Results

STAT 331. Martingale Central Limit Theorem and Related Results STAT 331 Martingale Central Limit Theorem and Related Results In this unit we discuss a version of the martingale central limit theorem, which states that under certain conditions, a sum of orthogonal

More information

Reflected Brownian Motion

Reflected Brownian Motion Chapter 6 Reflected Brownian Motion Often we encounter Diffusions in regions with boundary. If the process can reach the boundary from the interior in finite time with positive probability we need to decide

More information

Probability and Measure

Probability and Measure Part II Year 2018 2017 2016 2015 2014 2013 2012 2011 2010 2009 2008 2007 2006 2005 2018 84 Paper 4, Section II 26J Let (X, A) be a measurable space. Let T : X X be a measurable map, and µ a probability

More information

Dynkin (λ-) and π-systems; monotone classes of sets, and of functions with some examples of application (mainly of a probabilistic flavor)

Dynkin (λ-) and π-systems; monotone classes of sets, and of functions with some examples of application (mainly of a probabilistic flavor) Dynkin (λ-) and π-systems; monotone classes of sets, and of functions with some examples of application (mainly of a probabilistic flavor) Matija Vidmar February 7, 2018 1 Dynkin and π-systems Some basic

More information

On pathwise stochastic integration

On pathwise stochastic integration On pathwise stochastic integration Rafa l Marcin Lochowski Afican Institute for Mathematical Sciences, Warsaw School of Economics UWC seminar Rafa l Marcin Lochowski (AIMS, WSE) On pathwise stochastic

More information

JUSTIN HARTMANN. F n Σ.

JUSTIN HARTMANN. F n Σ. BROWNIAN MOTION JUSTIN HARTMANN Abstract. This paper begins to explore a rigorous introduction to probability theory using ideas from algebra, measure theory, and other areas. We start with a basic explanation

More information

Risk-Minimality and Orthogonality of Martingales

Risk-Minimality and Orthogonality of Martingales Risk-Minimality and Orthogonality of Martingales Martin Schweizer Universität Bonn Institut für Angewandte Mathematik Wegelerstraße 6 D 53 Bonn 1 (Stochastics and Stochastics Reports 3 (199, 123 131 2

More information

Stochastic integration. P.J.C. Spreij

Stochastic integration. P.J.C. Spreij Stochastic integration P.J.C. Spreij this version: April 22, 29 Contents 1 Stochastic processes 1 1.1 General theory............................... 1 1.2 Stopping times...............................

More information

SUMMARY OF RESULTS ON PATH SPACES AND CONVERGENCE IN DISTRIBUTION FOR STOCHASTIC PROCESSES

SUMMARY OF RESULTS ON PATH SPACES AND CONVERGENCE IN DISTRIBUTION FOR STOCHASTIC PROCESSES SUMMARY OF RESULTS ON PATH SPACES AND CONVERGENCE IN DISTRIBUTION FOR STOCHASTIC PROCESSES RUTH J. WILLIAMS October 2, 2017 Department of Mathematics, University of California, San Diego, 9500 Gilman Drive,

More information

Part II Probability and Measure

Part II Probability and Measure Part II Probability and Measure Theorems Based on lectures by J. Miller Notes taken by Dexter Chua Michaelmas 2016 These notes are not endorsed by the lecturers, and I have modified them (often significantly)

More information

Elements of Probability Theory

Elements of Probability Theory Elements of Probability Theory CHUNG-MING KUAN Department of Finance National Taiwan University December 5, 2009 C.-M. Kuan (National Taiwan Univ.) Elements of Probability Theory December 5, 2009 1 / 58

More information

Inference for Stochastic Processes

Inference for Stochastic Processes Inference for Stochastic Processes Robert L. Wolpert Revised: June 19, 005 Introduction A stochastic process is a family {X t } of real-valued random variables, all defined on the same probability space

More information

3 (Due ). Let A X consist of points (x, y) such that either x or y is a rational number. Is A measurable? What is its Lebesgue measure?

3 (Due ). Let A X consist of points (x, y) such that either x or y is a rational number. Is A measurable? What is its Lebesgue measure? MA 645-4A (Real Analysis), Dr. Chernov Homework assignment 1 (Due ). Show that the open disk x 2 + y 2 < 1 is a countable union of planar elementary sets. Show that the closed disk x 2 + y 2 1 is a countable

More information

STOCHASTIC CALCULUS JASON MILLER AND VITTORIA SILVESTRI

STOCHASTIC CALCULUS JASON MILLER AND VITTORIA SILVESTRI STOCHASTIC CALCULUS JASON MILLER AND VITTORIA SILVESTRI Contents Preface 1 1. Introduction 1 2. Preliminaries 4 3. Local martingales 1 4. The stochastic integral 16 5. Stochastic calculus 36 6. Applications

More information

Brownian Motion and Stochastic Calculus

Brownian Motion and Stochastic Calculus ETHZ, Spring 17 D-MATH Prof Dr Martin Larsson Coordinator A Sepúlveda Brownian Motion and Stochastic Calculus Exercise sheet 6 Please hand in your solutions during exercise class or in your assistant s

More information

Selected Exercises on Expectations and Some Probability Inequalities

Selected Exercises on Expectations and Some Probability Inequalities Selected Exercises on Expectations and Some Probability Inequalities # If E(X 2 ) = and E X a > 0, then P( X λa) ( λ) 2 a 2 for 0 < λ

More information

Jump Processes. Richard F. Bass

Jump Processes. Richard F. Bass Jump Processes Richard F. Bass ii c Copyright 214 Richard F. Bass Contents 1 Poisson processes 1 1.1 Definitions............................. 1 1.2 Stopping times.......................... 3 1.3 Markov

More information

Formulas for probability theory and linear models SF2941

Formulas for probability theory and linear models SF2941 Formulas for probability theory and linear models SF2941 These pages + Appendix 2 of Gut) are permitted as assistance at the exam. 11 maj 2008 Selected formulae of probability Bivariate probability Transforms

More information

A Concise Course on Stochastic Partial Differential Equations

A Concise Course on Stochastic Partial Differential Equations A Concise Course on Stochastic Partial Differential Equations Michael Röckner Reference: C. Prevot, M. Röckner: Springer LN in Math. 1905, Berlin (2007) And see the references therein for the original

More information

Probability Theory I: Syllabus and Exercise

Probability Theory I: Syllabus and Exercise Probability Theory I: Syllabus and Exercise Narn-Rueih Shieh **Copyright Reserved** This course is suitable for those who have taken Basic Probability; some knowledge of Real Analysis is recommended( will

More information

A NOTE ON STOCHASTIC INTEGRALS AS L 2 -CURVES

A NOTE ON STOCHASTIC INTEGRALS AS L 2 -CURVES A NOTE ON STOCHASTIC INTEGRALS AS L 2 -CURVES STEFAN TAPPE Abstract. In a work of van Gaans (25a) stochastic integrals are regarded as L 2 -curves. In Filipović and Tappe (28) we have shown the connection

More information

FOUNDATIONS OF MARTINGALE THEORY AND STOCHASTIC CALCULUS FROM A FINANCE PERSPECTIVE

FOUNDATIONS OF MARTINGALE THEORY AND STOCHASTIC CALCULUS FROM A FINANCE PERSPECTIVE FOUNDATIONS OF MARTINGALE THEORY AND STOCHASTIC CALCULUS FROM A FINANCE PERSPECTIVE JOSEF TEICHMANN 1. Introduction The language of mathematical Finance allows to express many results of martingale theory

More information

Stochastic Calculus and Black-Scholes Theory MTH772P Exercises Sheet 1

Stochastic Calculus and Black-Scholes Theory MTH772P Exercises Sheet 1 Stochastic Calculus and Black-Scholes Theory MTH772P Exercises Sheet. For ξ, ξ 2, i.i.d. with P(ξ i = ± = /2 define the discrete-time random walk W =, W n = ξ +... + ξ n. (i Formulate and prove the property

More information

The concentration of a drug in blood. Exponential decay. Different realizations. Exponential decay with noise. dc(t) dt.

The concentration of a drug in blood. Exponential decay. Different realizations. Exponential decay with noise. dc(t) dt. The concentration of a drug in blood Exponential decay C12 concentration 2 4 6 8 1 C12 concentration 2 4 6 8 1 dc(t) dt = µc(t) C(t) = C()e µt 2 4 6 8 1 12 time in minutes 2 4 6 8 1 12 time in minutes

More information

Stochastic Analysis. King s College London Version 1.4 October Markus Riedle

Stochastic Analysis. King s College London Version 1.4 October Markus Riedle Stochastic Analysis King s College London Version 1.4 October 215 Markus Riedle Preface These notes are for a course on Stochastic Analysis at King s College London. Given the limited time and diverse

More information

I. ANALYSIS; PROBABILITY

I. ANALYSIS; PROBABILITY ma414l1.tex Lecture 1. 12.1.2012 I. NLYSIS; PROBBILITY 1. Lebesgue Measure and Integral We recall Lebesgue measure (M411 Probability and Measure) λ: defined on intervals (a, b] by λ((a, b]) := b a (so

More information

Beyond the color of the noise: what is memory in random phenomena?

Beyond the color of the noise: what is memory in random phenomena? Beyond the color of the noise: what is memory in random phenomena? Gennady Samorodnitsky Cornell University September 19, 2014 Randomness means lack of pattern or predictability in events according to

More information

Probability Theory. Richard F. Bass

Probability Theory. Richard F. Bass Probability Theory Richard F. Bass ii c Copyright 2014 Richard F. Bass Contents 1 Basic notions 1 1.1 A few definitions from measure theory............. 1 1.2 Definitions............................. 2

More information

Measure and integration

Measure and integration Chapter 5 Measure and integration In calculus you have learned how to calculate the size of different kinds of sets: the length of a curve, the area of a region or a surface, the volume or mass of a solid.

More information

Weak convergence and large deviation theory

Weak convergence and large deviation theory First Prev Next Go To Go Back Full Screen Close Quit 1 Weak convergence and large deviation theory Large deviation principle Convergence in distribution The Bryc-Varadhan theorem Tightness and Prohorov

More information

Lecture 4: Introduction to stochastic processes and stochastic calculus

Lecture 4: Introduction to stochastic processes and stochastic calculus Lecture 4: Introduction to stochastic processes and stochastic calculus Cédric Archambeau Centre for Computational Statistics and Machine Learning Department of Computer Science University College London

More information

An Introduction to Malliavin calculus and its applications

An Introduction to Malliavin calculus and its applications An Introduction to Malliavin calculus and its applications Lecture 3: Clark-Ocone formula David Nualart Department of Mathematics Kansas University University of Wyoming Summer School 214 David Nualart

More information

P (A G) dp G P (A G)

P (A G) dp G P (A G) First homework assignment. Due at 12:15 on 22 September 2016. Homework 1. We roll two dices. X is the result of one of them and Z the sum of the results. Find E [X Z. Homework 2. Let X be a r.v.. Assume

More information

25.1 Ergodicity and Metric Transitivity

25.1 Ergodicity and Metric Transitivity Chapter 25 Ergodicity This lecture explains what it means for a process to be ergodic or metrically transitive, gives a few characterizes of these properties (especially for AMS processes), and deduces

More information

n [ F (b j ) F (a j ) ], n j=1(a j, b j ] E (4.1)

n [ F (b j ) F (a j ) ], n j=1(a j, b j ] E (4.1) 1.4. CONSTRUCTION OF LEBESGUE-STIELTJES MEASURES In this section we shall put to use the Carathéodory-Hahn theory, in order to construct measures with certain desirable properties first on the real line

More information

Homework # , Spring Due 14 May Convergence of the empirical CDF, uniform samples

Homework # , Spring Due 14 May Convergence of the empirical CDF, uniform samples Homework #3 36-754, Spring 27 Due 14 May 27 1 Convergence of the empirical CDF, uniform samples In this problem and the next, X i are IID samples on the real line, with cumulative distribution function

More information

for all f satisfying E[ f(x) ] <.

for all f satisfying E[ f(x) ] <. . Let (Ω, F, P ) be a probability space and D be a sub-σ-algebra of F. An (H, H)-valued random variable X is independent of D if and only if P ({X Γ} D) = P {X Γ}P (D) for all Γ H and D D. Prove that if

More information