Martingale Problems. Abhay G. Bhatt Theoretical Statistics and Mathematics Unit Indian Statistical Institute, Delhi
|
|
- Augusta Gibbs
- 5 years ago
- Views:
Transcription
1 s Abhay G. Bhatt Theoretical Statistics and Mathematics Unit Indian Statistical Institute, Delhi Lectures on Probability and Stochastic Processes III Indian Statistical Institute, Kolkata November 2008
2 Outline Introduction Associated Semigroups and Generators 1 Introduction Associated Semigroups and Generators 2 3 Time-Inhomogeneous Jump Perturbations
3 Introduction Associated Semigroups and Generators Let X be a E-valued process. X satisfies the Markov property if P (X t+r Γ σ(x s : 0 s t)) = P (X t+r Γ X t ) A function P(t, x, Γ) is called a transition function if P(t, x, ) is a probability measure for all (t, x) [0, ) E P(0, x, ) = δ x for all x E P(,, Γ) is measurable for all Γ E (Interpretation) P (X t Γ X 0 = x) = P(t, x, Γ) X is a Markov process if it admits a transition function so that ( ) P X t+s Γ Ft X = P(s, X t, Γ) t, s 0, Γ E Equivalently [ ] E f (X t+s ) Ft X = f (y)p(s, X t, dy) f C b (E)
4 Associated Semigroups Introduction Associated Semigroups and Generators Define T t f (x) = f (y)p(t, x, dy) = E [f (X t ) X 0 = x] T t T s = T t+s T t f 0 whenever f 0 T t f f [ E f (X t+s ) Ft X semigroup property positivity contraction ] = T s f (X t )
5 Generator Introduction Associated Semigroups and Generators A contraction semigroup {T t } satisfying lim T t f = T 0 f = f t 0 is called a Strongly continuous contraction semigroup The Generator L of a strongly continuous contraction semigroup is defined as follows. D(L) = T t f f = { f : lim t 0 T t f f t T t f f Lf = lim. t 0 t t 0 exists }. T s Lfds f D(L) (1)
6 Introduction Associated Semigroups and Generators Proposition 1 Let X, P, (T t ) and L be as above. Then X is a solution of the martingale problem for L. Proof. Fix f D(L) and let M t = f (X t ) t 0 Lf (X s)ds. [ ] [ ] t+s [ ] E M t+s Ft X = E f (X t+s ) Ft X E Lf (X u ) Ft X du 0 t+s t = T s f (X t ) T u t Lf (X t )du Lf (X u )du = T s f (X t ) t s 0 t T u Lf (X t )du Lf (X u )du 0 t = f (X t ) Lf (X u )du= M t (using (1)) 0 0
7 Outline 1 Introduction Associated Semigroups and Generators 2 3 Time-Inhomogeneous Jump Perturbations
8 Finite Dimensional Distributions Lemma 1 A process X is a solution to the martingale problem for A if and only if [ ( tn+1 ) n ] E f (X ) f (X t tn+1 n) Af (X s )ds h k (X tk ) = 0 (2) t n k=1 for all f D(A), 0 t 1 < t 2 <... < t n+1, h 1, h 2,..., h n B(E), and n 1. Thus, (being a) solution of the martingale problem is a finite dimensional property. Thus if X is a solution and Y is a modification of X, then Y is also a solution.
9 The space D([0, ), E) D([0, ), E) - the space of all E valued functions on [0, ) which are right continuous and have left limits Skorokhod topology on D([0, ), E) D([0, ), E) - complete, separable, metric space S E, the Borel σ-field on D([0, ), E). θ t (ω) = ω t - co-ordinate process r.c.l.l. process - process taking values in D([0, ), E)
10 r.c.l.l. Solutions Definition 2.1 A probability measure P P(D([0, ), E)) is solution of the martingale problem for (A, µ) if there exists a D([0, ), E)- valued process X with L(X ) = P and such that X is a solution to the martingale problem for (A, µ) Equivalently, P P(D([0, ), E)) is a solution if θ defined on (D([0, ), E), S E, P) is a solution For a r.c.l.l. process X defined on some (Ω, F, P), we will use the dual terminology X is a solution P X 1 is a solution
11 Well-posedness - Definitions Definition 2.2 The martingale problem for (A, µ) is well - posed in a class C of processes if there exists a solution X C of the martingale problem for (A, µ) and if Y C is also a solution to the martingale problem for (A, µ), then X and Y have the same finite dimensional distributions. i.e. uniqueness holds When C is the class of all measurable processes then we just say that the martingale problem is well - posed. Definition 2.3 The martingale problem for A is well - posed in C if the martingale problem for (A, µ) is well-posed for all µ P(E).
12 Well-posedness in D([0, ), E) finite dimensional distributions characterize the probability measures on D([0, ), E) Definition 2.4 The D([0, ), E) - martingale problem for (A, µ) is well - posed if there exists a solution P P(D([0, ), E)) of the D([0, ), E) - martingale problem for (A, µ) and if Q is any solution to the D([0, ), E) - martingale problem for (A, µ) then P = Q.
13 Bounded-pointwise convergence Definition 2.5 Let f k, f B(E). f k converge bounded-ly and pointwise to f bp f k f if f k M and f k (x) f (x) for all x E. A class of functions U B(E) bp-closed if f k U, f k bp f implies f U. bp-closure(u) - the smallest class of functions in B(E) which contains U and is bp-closed. i.e. E 0 - a field; σ(e 0 ) = E H - class of all E 0 -simple functions Then bp-closure(h) = class of all bounded, E - measurable functions.
14 Separability condition Definition 2.6 The operator A satisfies the separability condition if There exists a countable subset {f n } D(A) such that bp closure({(f n, Af n ) : n 1}) {(f, Af ) : f D(A)}. Let A 0 = A {fn}, the restriction of A to {f n } Solution of martingale problem for A = solution of martingale problem for A 0
15 Separability condition Definition 2.6 The operator A satisfies the separability condition if There exists a countable subset {f n } D(A) such that bp closure({(f n, Af n ) : n 1}) {(f, Af ) : f D(A)}. Let A 0 = A {fn}, the restriction of A to {f n } Solution of martingale problem for A solution of martingale problem for A 0 from Lemma 1 Use Dominated convergence Theorem to show that the set of all {(g, Ag)} satisfying (2) is bp-closed
16 Markov Family of Solutions Theorem 1 Let A be an operator on C b (E) satisfying the separability condition. Suppose the D([0, ), E) - martingale problem for (A, δ x ) is well-posed for each x E. Then 1 x P x (C) is measurable for all C S E. 2 For all µ P(E), the D([0, ), E) - martingale problem for (A, µ) is well - posed, with the solution P µ given by P µ (C) = P x (C)µ(dx). 3 Under P µ, θ t is a Markov process with transition function E P(s, x, F ) = P x (θ s F ). (3)
17 Proof of (1) Choose M C b (E) - countable such that B(E) bp-closure(m). { Let H = η : tm+1 η(θ) = (f n (θ tm+1 ) f n (θ tm ) t m Af n (θ s )ds) m h k (θ tk ) k=1 } where h 1, h 2,..., h m M, 0 t 1 < t 2... < t m+1 Q H is countable Lemma 1 = M 1 = η H {P : ηdp = 0} is the set of solutions of the martingale problem for A. P ηdp is continuous. Hence M 1 is Borel set
18 Proof of (1) (Contd.) G : P(D([0, ), E)) P(E) G is continuous G(P) = P θ(0) 1. M = M 1 G 1 ({δ x : x E}) = {P x : x E} is Borel Well-posedness = G restricted to M is one-to-one mapping onto {δ x : x E}. G 1 : {δ x : x E} M is Borel G(P x ) = δ x. Hence δ x P x is measurable x P x = x δ x P x is measurable
19 Proof of (2) For F E P µ θ 1 0 (F ) = For η H, E D([0, ),E) P x θ 1 0 (F )µ(dx) = ηdp µ = E D([0, ),E) E δ x (F )µ(dx) = µ(f ). ηdp x µ(dx) = 0. Hence P µ is a solution to the martingale problem for (A, µ). Let Q be another solution of the D([0, ), E) martingale problem for (A, µ) Let Q ω be the regular conditional probability of Q given θ 0.
20 Proof of (2) (Contd.) Fix η H, h C b (E). Define η (θ) = η(θ)h(θ 0 ). η H. Thus E Q [η(θ)h(θ 0 )] = E Q [η ] = 0. Since this holds for all h C b (E), E Qω [η] = E Q [η θ 0 ] = 0 a.s. - Q. Since H is countable, ONE Q-null set N 0 satisfying E Qω [η] = 0 ω N 0 Q ω is a solution of the martingale problem for A initial distribution δ θ0 (ω). Well - posedness implies Hence Q = P µ. Q ω = P θ0 (ω) a.s.[q]
21 Proof of (3) Fix s. Let θ t = θ t+s. Let Q ω be the regular conditional probability distribution of θ (under P x ) given F s. Q ω is a solution to the martingale problem for (A, δ θs(ω)). Well-posedness = Q ω(θ t F ) = P(t, θ s (ω), F ) (See (3)) Hence for f B(E), ] E Px f (θ t+s ) = E [E Px Px [f (θ t+s ) F s ] [ ] = E Px f (y)p(t, θ s ( ), dy) E = f (y 2 )P(t, y 1, dy 2 )P(s, x, dy 1 ). E E P(s + t, x, F ) = P x (θ t+s F ) = E P(t, y, F )P(s, x, dy)
22 One-dimensional equality Theorem 2 Suppose that for each µ P(E), any two solutions X and Y (defined respectively on (Ω 1, F 1, P 1 ) and (Ω 2, F 2, P 2 )) of the martingale problem for (A, µ) have the same one-dimensional distributions. Then X and Y have the same finite dimensional distributions, i.e. the martingale problem is well - posed. Proof. To show [ m ] [ m ] E P 1 f k (X tk ) = E P 2 f k (Y tk ) (4) k=1 k=1 for all 0 t 1 < t 2 <... t m, f 1, f 2,..., f m B(E) and m 1.
23 Induction argument Case: m = 1 - true by hypothesis Assume that the Induction hypothesis (4) is true for m = n Fix 0 t 1 < t 2 <... t n, f 1, f 2,..., f n B(E), f k > 0. Define Q 1 (F 1 ) = EP 1 [I F1 n k=1 f k(x tk )] E P 1 [ n k=1 f k(x tk )] Q 2 (F 2 ) = EP 2 [I F2 n k=1 f k(y tk )] E P 2 [ n k=1 f k(y tk )] F 1 F 1 F 2 F 2 Let X t = X tn+t, Ỹ t = Y tn+t. Fix 0 s 1 < s 2 <..., s m+1 = t, h 1, h 2,..., h m B(E) and f D(A).
24 Induction argument (Contd.) E P 1 η(θ) = [ η(x tn+ ) ( sm+1 ) m f (θ sm+1 ) f (θ sm ) Af (θ s )ds h k (θ tk ) s m k=1 ] n f k (X tk ) = k=1 E P 1[ ( f (X sm+1 +t n ) f (X sm+t n ) m h j (X tn+sj ) j=1 tn+s m+1 t n+s m ) Af (X u )du ] n f k (X tk ) = 0. k=1
25 Induction argument (Contd.) Hence E Q 1 [η( X )] = EP 1 [η(x n tn+ ) k=1 f k(x tk )] E P 1 [ n k=1 f = 0. k(x tk )] Similarly E Q 2 [η(ỹ )] = 0. X and Ỹ are solutions of the martingale problems for (A, L( X 0 )) and (A, L(Ỹ0)) respectively. E Q 1 [f ( X 0 )] = EP 1 [f (X tn ) n k=1 f k(x tk )] E P 1 [ n k=1 f k(x tk )] = EP 2 [f (Y tn ) n k=1 f k(y tk )] E P 2 [ n k=1 f = E Q 2 [f (Ỹ 0 )] f B(E). k(y tk )] This equality follows from induction hypothesis for m = n
26 Induction argument (Contd.) Hence X and Ỹ have the same initial distribution. One-dimensional uniqueness implies E Q 1 [f ( X t )] = E Q 2 [f (Ỹ t )] t 0, f B(E). E P 1 [f (X tn+t) n f k (X tk )] = E P 2 [f (X tn+t) k=1 Induction Hypothesis (4) is true for m = n + 1 set t n+1 = t n + t n f k (X tk )] k=1
27 Semigroup associated with the Suppose A satisfies the conditions of Theorem 1. Associate the Markov semigroup (T t ) t 0 with A - T t f (x) = f (y)p(t, x, dy) The following theorem can be proved exactly as the previous one. E
28 Strong Markov Property Theorem 3 Suppose that the D([0, ), E)- martingale problem for A is well - posed with associated semigroup T t Let X, defined on (Ω, F, P), be a solution of the martingale problem for A (with respect to (G t ) t 0 ). Let τ be a finite stop time. Then for f B(E), t 0, In particular E[f (X τ+t ) G τ ] = T t f (X τ ) P((X τ+t Γ) G τ ) = P(t, X τ, Γ) Γ E
29 r.c.l.l. modification Definition 2.7 Let D be a class of functions on E 1 D is measure determining if fdp = fdq forall f D implies P = Q. 2 D separates points in E if x y g D such that g(x) g(y). Theorem 4 Let E be a compact metric space. Let A be an operator on C(E) such that D(A) is measure determining and contains a countable subset that separates points in E. Let X, defined on (Ω, F, P), be a solution to the martingale problem for A. Then X has a modification with sample paths in D([0, ), E).
30 Proof Proof. Let {g k : k 1} D(A) separate points in E. Define t M k (t) = g k (X t ) Ag k (X s )ds 0 M k is a martingale for all k. Then for all t lim s t s Q M k (s), lim M k (s) exist a.s. s t s Q Hence Ω Ω with P(Ω ) = 1 and lim s t s Q g k (X s (ω)), lim g k (X s (ω)) exist ω Ω, t 0, k 1 s t s Q
31 Proof (contd.) Fix t 0, {s n } Q with s n > t, lim n s n = t and ω Ω. Since E is compact, a subsequence {s ni } such that lim i X sni (ω) exists. Clearly ( ) g k lim X s ni (ω) i = lim g k (X s (ω)) k. s t s Q Since {g k : k 1} separate points in E, lim s t X s (ω) exists. s Q Similarly lim s t X s (ω) exists s Q Define Y t (ω) = lim X s (ω). s t s Q
32 Proof (contd.) For ω Ω, Y t (ω) is r.c.l.l. & Define Y suitably for ω Ω Yt (ω) = lim X s (ω) s t s Q Then Y has sample paths in D([0, ), E). Since X is a solution to the martingale problem for A, for f D(A), a measure determining set = X = Y a.s. E[f (Y t ) Ft X ] = lim E[f (X s ) Ft X ] = f (X t ). s t s Q
33 Outline Time-Inhomogeneous Jump Perturbations 1 Introduction Associated Semigroups and Generators 2 3 Time-Inhomogeneous Jump Perturbations
34 Definitions Time-Inhomogeneous Jump Perturbations For t 0, let (A t ) t 0 be linear operators on M(E) with a common domain D M(E). Definition 3.1 A measurable process X defined on some probability space (Ω, F, P) is a solution to the martingale problem for (A t ) t 0 with respect to a filtration (G t ) t 0 if for any f D t f (X t ) A s f (X s )ds 0 is a (G t ) - martingale. Let µ P(E). The martingale problem for ((A t ) t 0, µ) is well-posed if there exists an unique solution for the martingale problem
35 Space-Time Process Time-Inhomogeneous Jump Perturbations Let E 0 = [0, ) E. Let Xt 0 = (t, X t ) Define { } k D(A 0 ) = g(t, x) = h i (t)f i (x) h i Cc 1 ([0, )), f i D i=1 A 0 g(t, x) = k [f i (x) t h i (t) + h i (t)a t f i (x)] i=1 Theorem 5 X is a solution to the martingale problem for (A t ) t 0 if and only if X 0 is a solution to the martingale problem for A 0.
36 Time-Inhomogeneous Jump Perturbations Proof. Let X be a solution (with respect to a filtration (G t ) t 0 ) to the martingale problem for (A t ) t 0. Let fh D(A 0 ) For 0 < s < t, let g(t) = E[f (X t ) G s ]. Then g(t) g(s) = g(t)h(t) g(s)h(s) = = = t s t s t s t s E[A u f (X u ) G s ]du u [g(u)h(u)] du {h(u)e[a u f (X u ) G s ] + g(u) u h(u)} du E [ A 0 (fh)(x 0 u ) G s ] du.
37 Proof (Contd.) Time-Inhomogeneous Jump Perturbations fh(x 0 (t)) t 0 A0 fh(x 0 (s))ds is a martingale X 0 is a solution to the martingale problem for A 0. The converse follows by taking h = 1 on [0, T ], T > 0.
38 A more General Result Time-Inhomogeneous Jump Perturbations State spaces E 1 and E 2 Operators A 1 on M(E 1 ) and A 2 on M(E 2 ) Solutions X 1 and X 2 Define D(A) = {f 1 f 2 : f 1 D(A 1 ), f 2 D(A 2 )} A(f 1 f 2 ) = (A 1 f 1 )f 2 + f 1 (A 2 f 2 ) (X 1, X 2 ) is a solution of the martingale problem for A Theorem 6 Suppose uniqueness holds for the martingale problem for A 1, A 2. Then uniqueness holds for the martingale problem for A.
39 A perturbed operator Time-Inhomogeneous Jump Perturbations Let A be an operator with D(A) C b (E). Let λ > 0 and let η(x, Γ) be a transition function on E E. Let Bf (x) = λ (f (y) f (x))η(x, dy) f B(E). E Theorem 7 Suppose that for every µ P(E), there exists a solution to the D([0, ), E) martingale problem for (A, µ). Then for every µ P(E) there exists a solution to the martingale problem for (A + B, µ).
40 Proof Time-Inhomogeneous Jump Perturbations Proof. For k 1, let Ω k = D([0, ), E), Ω 0 k = [0, ) Let Ω = k=1 Ω k Ω 0 k Let θ k and ξ k denote the co-ordinate random variables Borel σ-fields - F k, F 0 k Let F be the product σ-field on Ω. Let G k the σ-algebra generated by cylinder sets C 1 i=k+1 (Ω i Ω 0 i ), where C 1 F 1 F F k Fk 0. Let G k be the σ-algebra generated by k i=1 (Ω i Ω 0 i ) C 2, where C 2 F k F 0 k....
41 Perturbed Solution X Time-Inhomogeneous Jump Perturbations X evolves in E as a solution to the martingale problem for A till an exponentially distributed time with parameter λ which is independent of the past. At this time if the process is at x, it jumps to y with probability η(x, dy) and then continues evolving as a solution to the martingale problem for (A, δ y ). To put this in a mathematical framework, we consider that between the k th and the (k + 1) th jump (dictated by B), the process lies in Ω k. The k th copy of the exponential time is a random variable in Ω 0 k.
42 Proof (Contd.) Time-Inhomogeneous Jump Perturbations Let P x, P µ be solutions of the martingale problems for (A, δ x ), (A, µ) respectively. Let γ be the exponential distribution with parameter λ. Fix µ P(E). Define, for Γ 1 F 1,..., Γ k F k, F 1 F 0 1,..., F k F 0 k, P 1 (Γ 1 ) = P µ (Γ 1 ) ; P 0 1 (θ 1, F 1 ) = γ(f 1 ).. P k (θ 1, ξ 1,..., θ k 1, ξ k 1, Γ k ) = P 0 k (θ 1,..., ξ k 1, θ k, F k ) = γ(f k ) E P x (Γ k )η(θ k 1 (ξ k 1 ), dx)
43 Proof(Contd.) Time-Inhomogeneous Jump Perturbations P 1 P(Ω 1 ) and P1 0, P 2, P2 0,... are transition probability functions. an unique P on (Ω, F) satisfying For C G k and C G k+1 [ P(C C ) = E C ] P(C θ k+1 (0) = x)η(θ k (ξ k ), dx). Define τ 0 = 0, τ k = k i=1 ξ i N t = k for τ k t < τ k+1. Note that N is a Poisson process with parameter λ.
44 Proof(Contd.) Time-Inhomogeneous Jump Perturbations Define F t = F X t F N t. For f D(A) X t = θ k+1 (t τ k ), τ k t < τ k+1 f (θ k+1 ((t τ k ) τ k+1 τ k )) f (θ k+1 (0)) is an (F t ) t 0 martingale. Summing over k we get f (X t ) f (X 0 ) t is an (F t ) t 0 martingale. 0 (t τk ) τ k+1 τ k Af (θ k+1 (s τ k ))ds N(t) Af (X (s))ds (f (θ k+1 (0)) f (θ k (ξ k ))) k=1
45 Proof(Contd.) Time-Inhomogeneous Jump Perturbations Also, the following are (F t ) t 0 martingales. Hence N t k=1 t 0 ( ) f (θ k+1 (0)) f (y)(η(θ k (ξ k ), dy)) E E (f (y) f (X s )) η(x s, dy)d(n s λs) t f (X t ) f (X 0 ) (Af (X s ) + Bf (X s )) ds 0 is an (F t ) t 0 martingale.
An Ergodic Theorem for Fleming-Viot Models in Random Environments 1
An Ergodic Theorem for Fleming-Viot Models in Random Environments 1 Arash Jamshidpey 2 216 Abstract The Fleming-Viot (FV) process is a measure-valued diffusion that models the evolution of type frequencies
More informationFiltrations, Markov Processes and Martingales. Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition
Filtrations, Markov Processes and Martingales Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition David pplebaum Probability and Statistics Department,
More information{σ x >t}p x. (σ x >t)=e at.
3.11. EXERCISES 121 3.11 Exercises Exercise 3.1 Consider the Ornstein Uhlenbeck process in example 3.1.7(B). Show that the defined process is a Markov process which converges in distribution to an N(0,σ
More informationChapter 1. Poisson processes. 1.1 Definitions
Chapter 1 Poisson processes 1.1 Definitions Let (, F, P) be a probability space. A filtration is a collection of -fields F t contained in F such that F s F t whenever s
More information1. Stochastic Processes and filtrations
1. Stochastic Processes and 1. Stoch. pr., A stochastic process (X t ) t T is a collection of random variables on (Ω, F) with values in a measurable space (S, S), i.e., for all t, In our case X t : Ω S
More informationBrownian Motion. 1 Definition Brownian Motion Wiener measure... 3
Brownian Motion Contents 1 Definition 2 1.1 Brownian Motion................................. 2 1.2 Wiener measure.................................. 3 2 Construction 4 2.1 Gaussian process.................................
More informationPROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS
PROBABILITY: LIMIT THEOREMS II, SPRING 218. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please
More informationJump Processes. Richard F. Bass
Jump Processes Richard F. Bass ii c Copyright 214 Richard F. Bass Contents 1 Poisson processes 1 1.1 Definitions............................. 1 1.2 Stopping times.......................... 3 1.3 Markov
More informationWeak convergence and large deviation theory
First Prev Next Go To Go Back Full Screen Close Quit 1 Weak convergence and large deviation theory Large deviation principle Convergence in distribution The Bryc-Varadhan theorem Tightness and Prohorov
More informationSUMMARY OF RESULTS ON PATH SPACES AND CONVERGENCE IN DISTRIBUTION FOR STOCHASTIC PROCESSES
SUMMARY OF RESULTS ON PATH SPACES AND CONVERGENCE IN DISTRIBUTION FOR STOCHASTIC PROCESSES RUTH J. WILLIAMS October 2, 2017 Department of Mathematics, University of California, San Diego, 9500 Gilman Drive,
More informationDynkin (λ-) and π-systems; monotone classes of sets, and of functions with some examples of application (mainly of a probabilistic flavor)
Dynkin (λ-) and π-systems; monotone classes of sets, and of functions with some examples of application (mainly of a probabilistic flavor) Matija Vidmar February 7, 2018 1 Dynkin and π-systems Some basic
More informationLecture 17 Brownian motion as a Markov process
Lecture 17: Brownian motion as a Markov process 1 of 14 Course: Theory of Probability II Term: Spring 2015 Instructor: Gordan Zitkovic Lecture 17 Brownian motion as a Markov process Brownian motion is
More informationStochastic Processes II/ Wahrscheinlichkeitstheorie III. Lecture Notes
BMS Basic Course Stochastic Processes II/ Wahrscheinlichkeitstheorie III Michael Scheutzow Lecture Notes Technische Universität Berlin Sommersemester 218 preliminary version October 12th 218 Contents
More informationWiener Measure and Brownian Motion
Chapter 16 Wiener Measure and Brownian Motion Diffusion of particles is a product of their apparently random motion. The density u(t, x) of diffusing particles satisfies the diffusion equation (16.1) u
More informationInformation and Credit Risk
Information and Credit Risk M. L. Bedini Université de Bretagne Occidentale, Brest - Friedrich Schiller Universität, Jena Jena, March 2011 M. L. Bedini (Université de Bretagne Occidentale, Brest Information
More informationFunctional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals
Functional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals Noèlia Viles Cuadros BCAM- Basque Center of Applied Mathematics with Prof. Enrico
More informationExercises Measure Theoretic Probability
Exercises Measure Theoretic Probability 2002-2003 Week 1 1. Prove the folloing statements. (a) The intersection of an arbitrary family of d-systems is again a d- system. (b) The intersection of an arbitrary
More information4th Preparation Sheet - Solutions
Prof. Dr. Rainer Dahlhaus Probability Theory Summer term 017 4th Preparation Sheet - Solutions Remark: Throughout the exercise sheet we use the two equivalent definitions of separability of a metric space
More informationLecture 21 Representations of Martingales
Lecture 21: Representations of Martingales 1 of 11 Course: Theory of Probability II Term: Spring 215 Instructor: Gordan Zitkovic Lecture 21 Representations of Martingales Right-continuous inverses Let
More informationLecture 5. If we interpret the index n 0 as time, then a Markov chain simply requires that the future depends only on the present and not on the past.
1 Markov chain: definition Lecture 5 Definition 1.1 Markov chain] A sequence of random variables (X n ) n 0 taking values in a measurable state space (S, S) is called a (discrete time) Markov chain, if
More informationSOLUTIONS TO SOME PROBLEMS
23 FUNCTIONAL ANALYSIS Spring 23 SOLUTIONS TO SOME PROBLEMS Warning:These solutions may contain errors!! PREPARED BY SULEYMAN ULUSOY PROBLEM 1. Prove that a necessary and sufficient condition that the
More informationConvergence of Markov Processes. Amanda Turner University of Cambridge
Convergence of Markov Processes Amanda Turner University of Cambridge 1 Contents 1 Introduction 2 2 The Space D E [, 3 2.1 The Skorohod Topology................................ 3 3 Convergence of Probability
More informationLet (Ω, F) be a measureable space. A filtration in discrete time is a sequence of. F s F t
2.2 Filtrations Let (Ω, F) be a measureable space. A filtration in discrete time is a sequence of σ algebras {F t } such that F t F and F t F t+1 for all t = 0, 1,.... In continuous time, the second condition
More information6. Brownian Motion. Q(A) = P [ ω : x(, ω) A )
6. Brownian Motion. stochastic process can be thought of in one of many equivalent ways. We can begin with an underlying probability space (Ω, Σ, P) and a real valued stochastic process can be defined
More informationPROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS
PROBABILITY: LIMIT THEOREMS II, SPRING 15. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please
More informationA D VA N C E D P R O B A B I L - I T Y
A N D R E W T U L L O C H A D VA N C E D P R O B A B I L - I T Y T R I N I T Y C O L L E G E T H E U N I V E R S I T Y O F C A M B R I D G E Contents 1 Conditional Expectation 5 1.1 Discrete Case 6 1.2
More informationAn essay on the general theory of stochastic processes
Probability Surveys Vol. 3 (26) 345 412 ISSN: 1549-5787 DOI: 1.1214/1549578614 An essay on the general theory of stochastic processes Ashkan Nikeghbali ETHZ Departement Mathematik, Rämistrasse 11, HG G16
More informationRandom Process Lecture 1. Fundamentals of Probability
Random Process Lecture 1. Fundamentals of Probability Husheng Li Min Kao Department of Electrical Engineering and Computer Science University of Tennessee, Knoxville Spring, 2016 1/43 Outline 2/43 1 Syllabus
More informationStochastic integration. P.J.C. Spreij
Stochastic integration P.J.C. Spreij this version: April 22, 29 Contents 1 Stochastic processes 1 1.1 General theory............................... 1 1.2 Stopping times...............................
More informationELEMENTS OF PROBABILITY THEORY
ELEMENTS OF PROBABILITY THEORY Elements of Probability Theory A collection of subsets of a set Ω is called a σ algebra if it contains Ω and is closed under the operations of taking complements and countable
More informationTranslation Invariant Exclusion Processes (Book in Progress)
Translation Invariant Exclusion Processes (Book in Progress) c 2003 Timo Seppäläinen Department of Mathematics University of Wisconsin Madison, WI 53706-1388 December 11, 2008 Contents PART I Preliminaries
More informationBrownian Motion on Manifold
Brownian Motion on Manifold QI FENG Purdue University feng71@purdue.edu August 31, 2014 QI FENG (Purdue University) Brownian Motion on Manifold August 31, 2014 1 / 26 Overview 1 Extrinsic construction
More informationConvergence of Feller Processes
Chapter 15 Convergence of Feller Processes This chapter looks at the convergence of sequences of Feller processes to a iting process. Section 15.1 lays some ground work concerning weak convergence of processes
More informationfor all f satisfying E[ f(x) ] <.
. Let (Ω, F, P ) be a probability space and D be a sub-σ-algebra of F. An (H, H)-valued random variable X is independent of D if and only if P ({X Γ} D) = P {X Γ}P (D) for all Γ H and D D. Prove that if
More informationBrownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539
Brownian motion Samy Tindel Purdue University Probability Theory 2 - MA 539 Mostly taken from Brownian Motion and Stochastic Calculus by I. Karatzas and S. Shreve Samy T. Brownian motion Probability Theory
More informationSummer Jump-Start Program for Analysis, 2012 Song-Ying Li
Summer Jump-Start Program for Analysis, 01 Song-Ying Li 1 Lecture 6: Uniformly continuity and sequence of functions 1.1 Uniform Continuity Definition 1.1 Let (X, d 1 ) and (Y, d ) are metric spaces and
More informationFundamental Inequalities, Convergence and the Optional Stopping Theorem for Continuous-Time Martingales
Fundamental Inequalities, Convergence and the Optional Stopping Theorem for Continuous-Time Martingales Prakash Balachandran Department of Mathematics Duke University April 2, 2008 1 Review of Discrete-Time
More informationStochastic Calculus February 11, / 33
Martingale Transform M n martingale with respect to F n, n =, 1, 2,... σ n F n (σ M) n = n 1 i= σ i(m i+1 M i ) is a Martingale E[(σ M) n F n 1 ] n 1 = E[ σ i (M i+1 M i ) F n 1 ] i= n 2 = σ i (M i+1 M
More informationErgodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.
Ergodic Theorems Samy Tindel Purdue University Probability Theory 2 - MA 539 Taken from Probability: Theory and examples by R. Durrett Samy T. Ergodic theorems Probability Theory 1 / 92 Outline 1 Definitions
More informationStochastic Processes. Winter Term Paolo Di Tella Technische Universität Dresden Institut für Stochastik
Stochastic Processes Winter Term 2016-2017 Paolo Di Tella Technische Universität Dresden Institut für Stochastik Contents 1 Preliminaries 5 1.1 Uniform integrability.............................. 5 1.2
More informationStochastic Processes
Stochastic Processes A very simple introduction Péter Medvegyev 2009, January Medvegyev (CEU) Stochastic Processes 2009, January 1 / 54 Summary from measure theory De nition (X, A) is a measurable space
More informationLecture 19 L 2 -Stochastic integration
Lecture 19: L 2 -Stochastic integration 1 of 12 Course: Theory of Probability II Term: Spring 215 Instructor: Gordan Zitkovic Lecture 19 L 2 -Stochastic integration The stochastic integral for processes
More informationProperties of an infinite dimensional EDS system : the Muller s ratchet
Properties of an infinite dimensional EDS system : the Muller s ratchet LATP June 5, 2011 A ratchet source : wikipedia Plan 1 Introduction : The model of Haigh 2 3 Hypothesis (Biological) : The population
More informationReflected Brownian Motion
Chapter 6 Reflected Brownian Motion Often we encounter Diffusions in regions with boundary. If the process can reach the boundary from the interior in finite time with positive probability we need to decide
More informationp 1 ( Y p dp) 1/p ( X p dp) 1 1 p
Doob s inequality Let X(t) be a right continuous submartingale with respect to F(t), t 1 P(sup s t X(s) λ) 1 λ {sup s t X(s) λ} X + (t)dp 2 For 1 < p
More informationFeller Processes and Semigroups
Stat25B: Probability Theory (Spring 23) Lecture: 27 Feller Processes and Semigroups Lecturer: Rui Dong Scribe: Rui Dong ruidong@stat.berkeley.edu For convenience, we can have a look at the list of materials
More informationGeneralized Gaussian Bridges of Prediction-Invertible Processes
Generalized Gaussian Bridges of Prediction-Invertible Processes Tommi Sottinen 1 and Adil Yazigi University of Vaasa, Finland Modern Stochastics: Theory and Applications III September 1, 212, Kyiv, Ukraine
More informationExistence Results for Multivalued Semilinear Functional Differential Equations
E extracta mathematicae Vol. 18, Núm. 1, 1 12 (23) Existence Results for Multivalued Semilinear Functional Differential Equations M. Benchohra, S.K. Ntouyas Department of Mathematics, University of Sidi
More informationOn semilinear elliptic equations with measure data
On semilinear elliptic equations with measure data Andrzej Rozkosz (joint work with T. Klimsiak) Nicolaus Copernicus University (Toruń, Poland) Controlled Deterministic and Stochastic Systems Iasi, July
More informationNonlinear representation, backward SDEs, and application to the Principal-Agent problem
Nonlinear representation, backward SDEs, and application to the Principal-Agent problem Ecole Polytechnique, France April 4, 218 Outline The Principal-Agent problem Formulation 1 The Principal-Agent problem
More informationMATH 6605: SUMMARY LECTURE NOTES
MATH 6605: SUMMARY LECTURE NOTES These notes summarize the lectures on weak convergence of stochastic processes. If you see any typos, please let me know. 1. Construction of Stochastic rocesses A stochastic
More informationDo stochastic processes exist?
Project 1 Do stochastic processes exist? If you wish to take this course for credit, you should keep a notebook that contains detailed proofs of the results sketched in my handouts. You may consult any
More informationA PECULIAR COIN-TOSSING MODEL
A PECULIAR COIN-TOSSING MODEL EDWARD J. GREEN 1. Coin tossing according to de Finetti A coin is drawn at random from a finite set of coins. Each coin generates an i.i.d. sequence of outcomes (heads or
More informationLectures 16-17: Poisson Approximation. Using Lemma (2.4.3) with θ = 1 and then Lemma (2.4.4), which is valid when max m p n,m 1/2, we have
Lectures 16-17: Poisson Approximation 1. The Law of Rare Events Theorem 2.6.1: For each n 1, let X n,m, 1 m n be a collection of independent random variables with PX n,m = 1 = p n,m and PX n,m = = 1 p
More informationUniversal examples. Chapter The Bernoulli process
Chapter 1 Universal examples 1.1 The Bernoulli process First description: Bernoulli random variables Y i for i = 1, 2, 3,... independent with P [Y i = 1] = p and P [Y i = ] = 1 p. Second description: Binomial
More informationMean-field dual of cooperative reproduction
The mean-field dual of systems with cooperative reproduction joint with Tibor Mach (Prague) A. Sturm (Göttingen) Friday, July 6th, 2018 Poisson construction of Markov processes Let (X t ) t 0 be a continuous-time
More informationJUSTIN HARTMANN. F n Σ.
BROWNIAN MOTION JUSTIN HARTMANN Abstract. This paper begins to explore a rigorous introduction to probability theory using ideas from algebra, measure theory, and other areas. We start with a basic explanation
More informationMean-field SDE driven by a fractional BM. A related stochastic control problem
Mean-field SDE driven by a fractional BM. A related stochastic control problem Rainer Buckdahn, Université de Bretagne Occidentale, Brest Durham Symposium on Stochastic Analysis, July 1th to July 2th,
More informationADVANCED PROBABILITY: SOLUTIONS TO SHEET 1
ADVANCED PROBABILITY: SOLUTIONS TO SHEET 1 Last compiled: November 6, 213 1. Conditional expectation Exercise 1.1. To start with, note that P(X Y = P( c R : X > c, Y c or X c, Y > c = P( c Q : X > c, Y
More information(B(t i+1 ) B(t i )) 2
ltcc5.tex Week 5 29 October 213 Ch. V. ITÔ (STOCHASTIC) CALCULUS. WEAK CONVERGENCE. 1. Quadratic Variation. A partition π n of [, t] is a finite set of points t ni such that = t n < t n1
More informationOn continuous time contract theory
Ecole Polytechnique, France Journée de rentrée du CMAP, 3 octobre, 218 Outline 1 2 Semimartingale measures on the canonical space Random horizon 2nd order backward SDEs (Static) Principal-Agent Problem
More informationErgodic Properties of Markov Processes
Ergodic Properties of Markov Processes March 9, 2006 Martin Hairer Lecture given at The University of Warwick in Spring 2006 1 Introduction Markov processes describe the time-evolution of random systems
More informationApplications of Ito s Formula
CHAPTER 4 Applications of Ito s Formula In this chapter, we discuss several basic theorems in stochastic analysis. Their proofs are good examples of applications of Itô s formula. 1. Lévy s martingale
More informationHomework #6 : final examination Due on March 22nd : individual work
Université de ennes Année 28-29 Master 2ème Mathématiques Modèles stochastiques continus ou à sauts Homework #6 : final examination Due on March 22nd : individual work Exercise Warm-up : behaviour of characteristic
More informationChapter 6. Markov processes. 6.1 Introduction
Chapter 6 Markov processes 6.1 Introduction It is not uncommon for a Markov process to be defined as a sextuple (, F, F t,x t, t, P x ), and for additional notation (e.g.,,, S,P t,r,etc.) tobe introduced
More informationOPTIMAL SOLUTIONS TO STOCHASTIC DIFFERENTIAL INCLUSIONS
APPLICATIONES MATHEMATICAE 29,4 (22), pp. 387 398 Mariusz Michta (Zielona Góra) OPTIMAL SOLUTIONS TO STOCHASTIC DIFFERENTIAL INCLUSIONS Abstract. A martingale problem approach is used first to analyze
More informationAn Overview of the Martingale Representation Theorem
An Overview of the Martingale Representation Theorem Nuno Azevedo CEMAPRE - ISEG - UTL nazevedo@iseg.utl.pt September 3, 21 Nuno Azevedo (CEMAPRE - ISEG - UTL) LXDS Seminar September 3, 21 1 / 25 Background
More information9.2 Branching random walk and branching Brownian motions
168 CHAPTER 9. SPATIALLY STRUCTURED MODELS 9.2 Branching random walk and branching Brownian motions Branching random walks and branching diffusions have a long history. A general theory of branching Markov
More informationAn invariance result for Hammersley s process with sources and sinks
An invariance result for Hammersley s process with sources and sinks Piet Groeneboom Delft University of Technology, Vrije Universiteit, Amsterdam, and University of Washington, Seattle March 31, 26 Abstract
More informationVerona Course April Lecture 1. Review of probability
Verona Course April 215. Lecture 1. Review of probability Viorel Barbu Al.I. Cuza University of Iaşi and the Romanian Academy A probability space is a triple (Ω, F, P) where Ω is an abstract set, F is
More informationPROBABILITY THEORY II
Ruprecht-Karls-Universität Heidelberg Institut für Angewandte Mathematik Prof. Dr. Jan JOHANNES Outline of the lecture course PROBABILITY THEORY II Summer semester 2016 Preliminary version: April 21, 2016
More informationn E(X t T n = lim X s Tn = X s
Stochastic Calculus Example sheet - Lent 15 Michael Tehranchi Problem 1. Let X be a local martingale. Prove that X is a uniformly integrable martingale if and only X is of class D. Solution 1. If If direction:
More informationprocess on the hierarchical group
Intertwining of Markov processes and the contact process on the hierarchical group April 27, 2010 Outline Intertwining of Markov processes Outline Intertwining of Markov processes First passage times of
More informationZdzis law Brzeźniak and Tomasz Zastawniak
Basic Stochastic Processes by Zdzis law Brzeźniak and Tomasz Zastawniak Springer-Verlag, London 1999 Corrections in the 2nd printing Version: 21 May 2005 Page and line numbers refer to the 2nd printing
More informationNash Type Inequalities for Fractional Powers of Non-Negative Self-adjoint Operators. ( Wroclaw 2006) P.Maheux (Orléans. France)
Nash Type Inequalities for Fractional Powers of Non-Negative Self-adjoint Operators ( Wroclaw 006) P.Maheux (Orléans. France) joint work with A.Bendikov. European Network (HARP) (to appear in T.A.M.S)
More informationIntroduction and Preliminaries
Chapter 1 Introduction and Preliminaries This chapter serves two purposes. The first purpose is to prepare the readers for the more systematic development in later chapters of methods of real analysis
More informationReminder Notes for the Course on Measures on Topological Spaces
Reminder Notes for the Course on Measures on Topological Spaces T. C. Dorlas Dublin Institute for Advanced Studies School of Theoretical Physics 10 Burlington Road, Dublin 4, Ireland. Email: dorlas@stp.dias.ie
More informationReal Analysis Math 131AH Rudin, Chapter #1. Dominique Abdi
Real Analysis Math 3AH Rudin, Chapter # Dominique Abdi.. If r is rational (r 0) and x is irrational, prove that r + x and rx are irrational. Solution. Assume the contrary, that r+x and rx are rational.
More informationMATH 6337: Homework 8 Solutions
6.1. MATH 6337: Homework 8 Solutions (a) Let be a measurable subset of 2 such that for almost every x, {y : (x, y) } has -measure zero. Show that has measure zero and that for almost every y, {x : (x,
More informationThe Wiener Itô Chaos Expansion
1 The Wiener Itô Chaos Expansion The celebrated Wiener Itô chaos expansion is fundamental in stochastic analysis. In particular, it plays a crucial role in the Malliavin calculus as it is presented in
More informationBranching Processes II: Convergence of critical branching to Feller s CSB
Chapter 4 Branching Processes II: Convergence of critical branching to Feller s CSB Figure 4.1: Feller 4.1 Birth and Death Processes 4.1.1 Linear birth and death processes Branching processes can be studied
More informationWeak convergence and Brownian Motion. (telegram style notes) P.J.C. Spreij
Weak convergence and Brownian Motion (telegram style notes) P.J.C. Spreij this version: December 8, 2006 1 The space C[0, ) In this section we summarize some facts concerning the space C[0, ) of real
More informationMetric Spaces. Exercises Fall 2017 Lecturer: Viveka Erlandsson. Written by M.van den Berg
Metric Spaces Exercises Fall 2017 Lecturer: Viveka Erlandsson Written by M.van den Berg School of Mathematics University of Bristol BS8 1TW Bristol, UK 1 Exercises. 1. Let X be a non-empty set, and suppose
More informationBrownian Motion and Stochastic Calculus
ETHZ, Spring 17 D-MATH Prof Dr Martin Larsson Coordinator A Sepúlveda Brownian Motion and Stochastic Calculus Exercise sheet 6 Please hand in your solutions during exercise class or in your assistant s
More informationMAT 578 FUNCTIONAL ANALYSIS EXERCISES
MAT 578 FUNCTIONAL ANALYSIS EXERCISES JOHN QUIGG Exercise 1. Prove that if A is bounded in a topological vector space, then for every neighborhood V of 0 there exists c > 0 such that tv A for all t > c.
More informationMarkov Property and Ergodicity of the Nonlinear Filter
Markov Property and Ergodicity of the Nonlinear Filter A.G. Bhatt Indian Statistical Institute, New Delhi A. Budhiraja Department of Mathematics University of Notre Dame Notre Dame, IN 46656 R. L. Karandikar
More informationµ X (A) = P ( X 1 (A) )
1 STOCHASTIC PROCESSES This appendix provides a very basic introduction to the language of probability theory and stochastic processes. We assume the reader is familiar with the general measure and integration
More informationProduct measures, Tonelli s and Fubini s theorems For use in MAT4410, autumn 2017 Nadia S. Larsen. 17 November 2017.
Product measures, Tonelli s and Fubini s theorems For use in MAT4410, autumn 017 Nadia S. Larsen 17 November 017. 1. Construction of the product measure The purpose of these notes is to prove the main
More informationLecture 12. F o s, (1.1) F t := s>t
Lecture 12 1 Brownian motion: the Markov property Let C := C(0, ), R) be the space of continuous functions mapping from 0, ) to R, in which a Brownian motion (B t ) t 0 almost surely takes its value. Let
More information16 1 Basic Facts from Functional Analysis and Banach Lattices
16 1 Basic Facts from Functional Analysis and Banach Lattices 1.2.3 Banach Steinhaus Theorem Another fundamental theorem of functional analysis is the Banach Steinhaus theorem, or the Uniform Boundedness
More informationApproximating irregular SDEs via iterative Skorokhod embeddings
Approximating irregular SDEs via iterative Skorokhod embeddings Universität Duisburg-Essen, Essen, Germany Joint work with Stefan Ankirchner and Thomas Kruse Stochastic Analysis, Controlled Dynamical Systems
More informationIntegration on Measure Spaces
Chapter 3 Integration on Measure Spaces In this chapter we introduce the general notion of a measure on a space X, define the class of measurable functions, and define the integral, first on a class of
More informationElliptic Operators with Unbounded Coefficients
Elliptic Operators with Unbounded Coefficients Federica Gregorio Universitá degli Studi di Salerno 8th June 2018 joint work with S.E. Boutiah, A. Rhandi, C. Tacelli Motivation Consider the Stochastic Differential
More information7 Convergence in R d and in Metric Spaces
STA 711: Probability & Measure Theory Robert L. Wolpert 7 Convergence in R d and in Metric Spaces A sequence of elements a n of R d converges to a limit a if and only if, for each ǫ > 0, the sequence a
More informationSome Properties of NSFDEs
Chenggui Yuan (Swansea University) Some Properties of NSFDEs 1 / 41 Some Properties of NSFDEs Chenggui Yuan Swansea University Chenggui Yuan (Swansea University) Some Properties of NSFDEs 2 / 41 Outline
More informationProcesses with independent increments.
Dylan Gonzalez Arroyo Processes with independent increments. Master Thesis Thesis supervisor: Dr. F.M. Spieksma, Dr. O. van Gaans Date master exam: 29 January 2016 Mathematisch Instituut, Universiteit
More informationDistance-Divergence Inequalities
Distance-Divergence Inequalities Katalin Marton Alfréd Rényi Institute of Mathematics of the Hungarian Academy of Sciences Motivation To find a simple proof of the Blowing-up Lemma, proved by Ahlswede,
More informationInsert your Booktitle, Subtitle, Edition
C. Landim Insert your Booktitle, Subtitle, Edition SPIN Springer s internal project number, if known Monograph October 23, 2018 Springer Page: 1 job: book macro: svmono.cls date/time: 23-Oct-2018/15:27
More informationTyler Hofmeister. University of Calgary Mathematical and Computational Finance Laboratory
JUMP PROCESSES GENERALIZING STOCHASTIC INTEGRALS WITH JUMPS Tyler Hofmeister University of Calgary Mathematical and Computational Finance Laboratory Overview 1. General Method 2. Poisson Processes 3. Diffusion
More informationTHE SKOROKHOD OBLIQUE REFLECTION PROBLEM IN A CONVEX POLYHEDRON
GEORGIAN MATHEMATICAL JOURNAL: Vol. 3, No. 2, 1996, 153-176 THE SKOROKHOD OBLIQUE REFLECTION PROBLEM IN A CONVEX POLYHEDRON M. SHASHIASHVILI Abstract. The Skorokhod oblique reflection problem is studied
More information