Simulation of conditional diffusions via forward-reverse stochastic representations
|
|
- Jonah Lester
- 5 years ago
- Views:
Transcription
1 Weierstrass Institute for Applied Analysis and Stochastics Simulation of conditional diffusions via forward-reverse stochastic representations Christian Bayer and John Schoenmakers Numerical methods for stochastic differential equations UT Vienna, 2-4 September 2013 Mohrenstrasse Berlin Germany Tel
2 Outline 1 Introduction 2 The forward-reverse method for transition density estimation 3 A forward-reverse representation for conditional expectations 4 The forward-reverse estimator 5 Numerical examples Forward-reverse representation for conditional diffusions Page 2 (26)
3 Outline 1 Introduction 2 The forward-reverse method for transition density estimation 3 A forward-reverse representation for conditional expectations 4 The forward-reverse estimator 5 Numerical examples Forward-reverse representation for conditional diffusions Page 3 (26)
4 Conditional expectations Given a : [0,T] R d R d, σ : [0,T] R d R d m, a standard, m-dimensional Brownian motion B, consider dx(s) = a(s, X(s))ds + σ(s, X(s))dB(s), 0 s T Goal Given a grid D = {0 = s 0 < < s K+L+1 = T }, f : R (K+L)d R and x,y R d, compute E [ f (X(s 1 ),..., X(s K+L )) X(0) = x, X(T) = y ]. Forward-reverse representation for conditional diffusions Page 4 (26)
5 Conditional expectations Given a : [0,T] R d R d, σ : [0,T] R d R d m, a standard, m-dimensional Brownian motion B, consider dx(s) = a(s, X(s))ds + σ(s, X(s))dB(s), 0 s T Goal (extended) Given a grid D = {0 = s 0 < < s K+L+1 = T }, f : R (K+L)d R and A, B R d, compute E [ f (X(s 1 ),..., X(s K+L )) X(0) A, X(T) B ], A, B with positive measure or d -dimensional hyperplanes, 0 d d. Forward-reverse representation for conditional diffusions Page 4 (26)
6 Bridge simulation using a modified drift (Delyon and Hu 2006) ( dy(s) = a(s,y(s)) Y(s) y ) ds + σ(s,y(s))db(s), T s Y(0) = x SDE admits a unique solution Y on [0,T[ with lim s T Y(s) = y The law of Y on path-space is absolutely continuous w.r.t. the law of X conditioned on X(T) = y, X(0) = x. The Radon-Nikodym derivative is explicitly given (up to a constant) as an integral of Y(s), σ 1 (s,y(s)) and quadratic co-variations between them. Forward-reverse representation for conditional diffusions Page 5 (26)
7 Bridge simulation by time reversal (Bladt and Sørensen 2012) Dimension d = 1 X (1) t solution of SDE started at X (1) 0 = x X (2) t solution of SDE started at X (2) 0 = y τ inf {0 t T X (1) t = X (2) T t } Z t X (1) t, 0 t τ, Z t X (2) T t, τ t T on {τ T }. Theorem (Bladt and Sørensen 2009) The distribution of Z given {τ T } is equal to the distribution of a bridge process given that the bridge is hit by an independent realization of the SDE with initial distribution p(t, y, x)dx. Crucial hitting probability of bridge and time-reversed diffusion hard to estimate Forward-reverse representation for conditional diffusions Page 6 (26)
8 Bridge simulation by time reversal (Bladt and Sørensen 2012) Dimension d = 1 X (1) t solution of SDE started at X (1) 0 = x X (2) t solution of SDE started at X (2) 0 = y τ inf {0 t T X (1) t = X (2) T t } Z t X (1) t, 0 t τ, Z t X (2) T t, τ t T on {τ T }. Theorem (Bladt and Sørensen 2009) The distribution of Z given {τ T } is equal to the distribution of a bridge process given that the bridge is hit by an independent realization of the SDE with initial distribution p(t, y, x)dx. Crucial hitting probability of bridge and time-reversed diffusion hard to estimate Forward-reverse representation for conditional diffusions Page 6 (26)
9 Bridge simulation by time reversal (Bladt and Sørensen 2012) Dimension d = 1 X (1) t solution of SDE started at X (1) 0 = x X (2) t solution of SDE started at X (2) 0 = y τ inf {0 t T X (1) t = X (2) T t } Z t X (1) t, 0 t τ, Z t X (2) T t, τ t T on {τ T }. Theorem (Bladt and Sørensen 2009) The distribution of Z given {τ T } is equal to the distribution of a bridge process given that the bridge is hit by an independent realization of the SDE with initial distribution p(t, y, x)dx. Crucial hitting probability of bridge and time-reversed diffusion hard to estimate Forward-reverse representation for conditional diffusions Page 6 (26)
10 Outline 1 Introduction 2 The forward-reverse method for transition density estimation 3 A forward-reverse representation for conditional expectations 4 The forward-reverse estimator 5 Numerical examples Forward-reverse representation for conditional diffusions Page 7 (26)
11 Forward process Notation: X t,x (s) solution of SDE started at X t,x (t) = x, t < s Generator of the SDE: L t f (x) = f (x), a(t, x) d b i j (t, x) x i x j f (x), i, j=1 where b i j (x) = σ(t, x)σ(t, x) T Transition density p(t, x, T, y) Forward representation (Feynman Kac formula) u(t, x) = E [ f ( X t,x (T) )] = p(t, x,t,y) f (y)dy I( f ) t u(t, x) + L t u(t, x) = 0, u(t, x) = f (x) Forward-reverse representation for conditional diffusions Page 8 (26)
12 Forward process Notation: X t,x (s) solution of SDE started at X t,x (t) = x, t < s Generator of the SDE: L t f (x) = f (x), a(t, x) d b i j (t, x) x i x j f (x), i, j=1 where b i j (x) = σ(t, x)σ(t, x) T Transition density p(t, x, T, y) Forward representation (Feynman Kac formula) u(t, x) = E [ f ( X t,x (T) )] = p(t, x,t,y) f (y)dy I( f ) t u(t, x) + L t u(t, x) = 0, u(t, x) = f (x) Forward-reverse representation for conditional diffusions Page 8 (26)
13 Reverse process Consider: v(s,y) g(x)p(t, x, s,y)dx Forward-reverse representation for conditional diffusions Page 9 (26)
14 Reverse process Consider: v(s,y) g(x)p(t, x, s,y)dx Fokker-Planck equation: s p(t, x, s,y) = 1 2 d ( y i y j b i j (s,y)p(t, x, s,y) ) i, j=1 d i=1 y i ( a i (s,y)p(t, x, s,y) ) Forward-reverse representation for conditional diffusions Page 9 (26)
15 Reverse process Consider: v(s,y) g(x)p(t, x, s,y)dx Cauchy problem for v s v(s,y) = 1 2 d ( y i y j b i j (s,y)v(s,y) ) i, j=1 v(t,y) = g(y) d i=1 y i ( a i (s,y)v(s,y) ), Forward-reverse representation for conditional diffusions Page 9 (26)
16 Reverse process Consider: v(s,y) g(x)p(t, x, s,y)dx Cauchy problem for ṽ(s,y) v(t + t s,y) s ṽ(s,y) d y i y j ( b i j (s,y)ṽ(s,y) ) i, j=1 ṽ(t,y) = g(y), d i=1 y i (ãi (s,y)ṽ(s,y) ) = 0, where b(s,y) b(t + t s,y), ã(s,y) a(t + t s,y) Forward-reverse representation for conditional diffusions Page 9 (26)
17 Reverse process Consider: v(s,y) g(x)p(t, x, s,y)dx Cauchy problem for ṽ(s,y) v(t + t s,y) s ṽ(s,y) d d b i j (s,y) y i y jṽ(s,y) + α i (s,y) y iṽ(s,y) + c(y)ṽ(s,y) = 0, i, j=1 i=1 ṽ(t,y) = g(y), where c(s,y) = 1 2 d α i (s,y) y j b i j (y) ã i (s,y), i, j=1 j=1 d d y i y j b i j (s,y) y iã i (s,y) i=1 Forward-reverse representation for conditional diffusions Page 9 (26)
18 Reverse process Consider: v(s,y) g(x)p(t, x, s,y)dx Reverse representation (Feynman-Kac formula) I (g) v(t,y) = E [ g ( Y t,y (T) ) Y t,y,1 (T) ] dy(s) = α(s, Y(s))ds + σ(s, Y(s))dB(s), Y(t) = y, dy(s) = c(s, Y(s))Y(s)ds, Y(t) = 1 d α i (s,y) y j b i j (y) ã i (s,y), c(s,y) = 1 2 i, j=1 j=1 d d y i y j b i j (s,y) y iã i (s,y) i=1 Forward-reverse representation for conditional diffusions Page 9 (26)
19 Forward-reverse representation Theorem (Milstein, Schoenmakers, Spokoiny 2004) Choose X and (Y,Y) independent, t < t < T: E [ f ( X t,x (t ),Y t,y(t) ) Y t,y(t) ] = = p(t, x,t, x ) f (x,y )p(t,y,t,y)dx dy J( f ). Proof. Condition on X t,x (t ) and apply the reverse representation Integrate with respect to the law of X t,x (t ) Forward-reverse representation for conditional diffusions Page 10 (26)
20 Forward-reverse representation Theorem (Milstein, Schoenmakers, Spokoiny 2004) Choose X and (Y,Y) independent, t < t < T: E [ f ( X t,x (t ),Y t,y(t) ) Y t,y(t) ] = = p(t, x,t, x ) f (x,y )p(t,y,t,y)dx dy J( f ). Proof. Condition on X t,x (t ) and apply the reverse representation Integrate with respect to the law of X t,x (t ) Forward-reverse representation for conditional diffusions Page 10 (26)
21 Forward-reverse estimator for p(t, x, T, y) Formally inserting f (x,y ) = δ 0 (x y ) gives J( f ) = p(t, x,t,y) Use kernel f (x,y ) = ɛ d K ( ) x y with bandwidth ɛ > 0 Define estimator: ˆp N,M,ɛ 1 N M ɛ d Yt m MN,y (T)K Xt,x n (t ) Yt m,y (T) ɛ n=1 m=1 Theorem (Milstein, Schoenmakers, Spokoiny 2004) Assume that the coefficients of the SDE are C bounded and satisfy a uniform ellipticity (or uniform Hörmander) condition. ɛ If d 4, choose M = N, ɛ N = CN 1/4, then the MSE of ˆp N,N,ɛN is of order N 1. For d > 4, choose M = N and ɛ N = CN 2/(4+d), then the MSE of ˆp N,N,ɛN is of order N 8/(4+d). Forward-reverse representation for conditional diffusions Page 11 (26)
22 Forward-reverse estimator for p(t, x, T, y) Formally inserting f (x,y ) = δ 0 (x y ) gives J( f ) = p(t, x,t,y) Use kernel f (x,y ) = ɛ d K ( ) x y with bandwidth ɛ > 0 Define estimator: ˆp N,M,ɛ 1 N M ɛ d Yt m MN,y (T)K Xt,x n (t ) Yt m,y (T) ɛ n=1 m=1 Theorem (Milstein, Schoenmakers, Spokoiny 2004) Assume that the coefficients of the SDE are C bounded and satisfy a uniform ellipticity (or uniform Hörmander) condition. ɛ If d 4, choose M = N, ɛ N = CN 1/4, then the MSE of ˆp N,N,ɛN is of order N 1. For d > 4, choose M = N and ɛ N = CN 2/(4+d), then the MSE of ˆp N,N,ɛN is of order N 8/(4+d). Forward-reverse representation for conditional diffusions Page 11 (26)
23 Outline 1 Introduction 2 The forward-reverse method for transition density estimation 3 A forward-reverse representation for conditional expectations 4 The forward-reverse estimator 5 Numerical examples Forward-reverse representation for conditional diffusions Page 12 (26)
24 Generalization to multiple time steps Fix a grid 0 < t < t 1 < < t L := T Theorem (Generalization MSS 2004) Introduce the re-ordered grid t < ˆt 1 < < ˆt L = T defined by ˆt i T + t t L i, i = 1,..., L. Then (with t 0 = t and y L+1 = y) E [ f (Y t,y(t),y t,y(ˆt L 1 ),...,Y t,y(ˆt 1 ))Y t,y(t) ] = L f (y 1,y 2,...,y L ) p(t i 1,y i,t i,y i+1 )dy i. R d L i=1 Forward-reverse representation for conditional diffusions Page 13 (26)
25 Generalization to multiple time steps, cont. Key elements of the proof. Initially, we proved the autonomous case by direct induction on L first. Direct induction on L failed to work in the non-autonomous case due to the dependence of Y t,y on both t and T. For the non-autonomous case we lifted to autonomous in a standard wy while adding noise δdw for keeping non-degeneracy. Apply autonomous result and let δ 0 About five pages proof... Forward-reverse representation for conditional diffusions Page 14 (26)
26 Generalization to multiple time steps, cont. Key elements of the proof. Initially, we proved the autonomous case by direct induction on L first. Direct induction on L failed to work in the non-autonomous case due to the dependence of Y t,y on both t and T. For the non-autonomous case we lifted to autonomous in a standard wy while adding noise δdw for keeping non-degeneracy. Apply autonomous result and let δ 0 About five pages proof... Forward-reverse representation for conditional diffusions Page 14 (26)
27 Generalization to multiple time steps, cont. Key elements of the proof. Initially, we proved the autonomous case by direct induction on L first. Direct induction on L failed to work in the non-autonomous case due to the dependence of Y t,y on both t and T. For the non-autonomous case we lifted to autonomous in a standard wy while adding noise δdw for keeping non-degeneracy. Apply autonomous result and let δ 0 About five pages proof... Forward-reverse representation for conditional diffusions Page 14 (26)
28 Generalization to multiple time steps, cont. Key elements of the proof. Initially, we proved the autonomous case by direct induction on L first. Direct induction on L failed to work in the non-autonomous case due to the dependence of Y t,y on both t and T. For the non-autonomous case we lifted to autonomous in a standard wy while adding noise δdw for keeping non-degeneracy. Apply autonomous result and let δ 0 About five pages proof... Forward-reverse representation for conditional diffusions Page 14 (26)
29 Generalization to multiple time steps, cont. New proof. Get rid of the dependence of Y t,y on t by observing that Y t,y(s) = Y 0,y (s t ) =: Y y (s t ), Y t,y(s) = Y 0,y (s t ) =: Y y (s t ), t s T, and consequently E[ f (Y y;t (T t ))Y y;t (T t )] = p(t,y,t,y) f (y )dy, holds for any 0 < t < T in terms of one and the same reverse process Y y (s t ) Direct induction works... only two pages! Forward-reverse representation for conditional diffusions Page 15 (26)
30 Generalization to multiple time steps, cont. New proof. Get rid of the dependence of Y t,y on t by observing that Y t,y(s) = Y 0,y (s t ) =: Y y (s t ), Y t,y(s) = Y 0,y (s t ) =: Y y (s t ), t s T, and consequently E[ f (Y y;t (T t ))Y y;t (T t )] = p(t,y,t,y) f (y )dy, holds for any 0 < t < T in terms of one and the same reverse process Y y (s t ) Direct induction works... only two pages! Forward-reverse representation for conditional diffusions Page 15 (26)
31 Generalization to multiple time steps, cont. New proof. Get rid of the dependence of Y t,y on t by observing that Y t,y(s) = Y 0,y (s t ) =: Y y (s t ), Y t,y(s) = Y 0,y (s t ) =: Y y (s t ), t s T, and consequently E[ f (Y y;t (T t ))Y y;t (T t )] = p(t,y,t,y) f (y )dy, holds for any 0 < t < T in terms of one and the same reverse process Y y (s t ) Direct induction works... only two pages! Forward-reverse representation for conditional diffusions Page 15 (26)
32 A forward-reverse representation of the conditional expectation Consider a grid 0 = s 0 < < s K+L = T Choose t s K, rename t i s i+k, 0 i L Assume p(s 0, x,t,y) > 0 K : R d R, K(u)du = 1 Theorem Let ˆt i T + t t L i, X t = X s0,x(t), Y t = Y t,y(t), Y t = Y t,y(t), X and (Y,Y) independent, then E [ g(x s1,..., X sk+l 1 ) XT = y ] = 1 [ ( p(s 0, x,t,y) lime g(x s1,..., X t,yˆtl 1,...,Yˆt1 )ɛ d YT X t K ɛ 0 ɛ )Y T ]. Forward-reverse representation for conditional diffusions Page 16 (26)
33 Outline 1 Introduction 2 The forward-reverse method for transition density estimation 3 A forward-reverse representation for conditional expectations 4 The forward-reverse estimator 5 Numerical examples Forward-reverse representation for conditional diffusions Page 17 (26)
34 Assumptions Conditions on the diffusion process: Transition densities p(s, x,t,y) and q(s, x,t,y) of X and Y exist and α x β y p(s, x,t,y) C 1 ( C (t s) ν exp y x 2 ) 2 t s for multi-indices α + β 2 (and sim. for q). Moreover, p(s 0, x,t,y) > 0. Conditions on the kernel: K(v)dv = 1, vk(v)dv = 0 (second order) K(v) C exp ( α v 2+β), C,α,β 0, v R d Conditions on the function: g and its first and second derivatives are polynomially bounded Forward-reverse representation for conditional diffusions Page 18 (26)
35 Assumptions Conditions on the diffusion process: Transition densities p(s, x,t,y) and q(s, x,t,y) of X and Y exist and α x β y p(s, x,t,y) C 1 ( C (t s) ν exp y x 2 ) 2 t s for multi-indices α + β 2 (and sim. for q). Moreover, p(s 0, x,t,y) > 0. Conditions on the kernel: K(v)dv = 1, vk(v)dv = 0 (second order) K(v) C exp ( α v 2+β), C,α,β 0, v R d Conditions on the function: g and its first and second derivatives are polynomially bounded Forward-reverse representation for conditional diffusions Page 18 (26)
36 Assumptions Conditions on the diffusion process: Transition densities p(s, x,t,y) and q(s, x,t,y) of X and Y exist and α x β y p(s, x,t,y) C 1 ( C (t s) ν exp y x 2 ) 2 t s for multi-indices α + β 2 (and sim. for q). Moreover, p(s 0, x,t,y) > 0. Conditions on the kernel: K(v)dv = 1, vk(v)dv = 0 (second order) K(v) C exp ( α v 2+β), C,α,β 0, v R d Conditions on the function: g and its first and second derivatives are polynomially bounded Forward-reverse representation for conditional diffusions Page 18 (26)
37 The forward-reverse estimator for the conditional expectation Stochastic representation: H E [ g(x s1,..., X tl 1 ) XT = y ] = [ ( lime g(x s1,..., X t,yˆtl 1,...,Yˆt1 )ɛ d YT X t K ɛ 0 ɛ )Y T ] /p(s 0, x,t,y) Forward-reverse representation for conditional diffusions Page 19 (26)
38 The forward-reverse estimator for the conditional expectation Stochastic representation: H E [ g(x s1,..., X tl 1 ) XT = y ] = [ ( lime g(x s1,..., X t,yˆtl 1,...,Yˆt1 )ɛ d YT X t K ɛ 0 ɛ Estimator: for (X n ) i.i.d., (Y m,y m ) i.i.d., Ĥ ɛ,m,n := )Y T ] /p(s 0, x,t,y) Nn=1 Mm=1 g ( X n s 1,..., X n s K,Y mˆt L 1,...,Ymˆt 1 ) K ( Y m T X n t Nn=1 ( Mm=1 Y m ) K T X n t ɛ YT m 1 1 NM ɛ d N n=1 Mm=1 K ( Y m T X n t ɛ ɛ ) Y m T ) Y m T >p/2, Forward-reverse representation for conditional diffusions Page 19 (26)
39 The forward-reverse estimator for the conditional expectation Estimator: for (X n ) i.i.d., (Y m,y m ) i.i.d., Ĥ ɛ,m,n := Nn=1 Mm=1 g ( X n s 1,..., X n s K,Y mˆt L 1,...,Ymˆt 1 ) K ( Y m T X n t Nn=1 ( Mm=1 Y m ) K T X n t ɛ YT m 1 1 NM ɛ d N n=1 Mm=1 K ( Y m T X n t ɛ ɛ ) Y m T ) Y m T >p/2, Theorem Assume p(s 0, x,t,y) > p > 0, choose M = N. [( ) For d 4, set ɛ N = CN 1/4 2 ], then E H ĤN,N,ɛN = O(N 1 ). For d > 4, set ɛ N = CN 2/(4+d), then [( ) 2 ] E H ĤN,N,ɛN = O(N 8/(4+d) ). Forward-reverse representation for conditional diffusions Page 19 (26)
40 Implementation Assume K has compact support in B r (0) Algorithm 1. Simulate N indep. trajectories (X n ) N n=1 started at X s 0 = x and (Y m ) N m=1 started at Y t = y on D [s 0,t ] and D [t,t], resp. 2. For fixed m {1,..., N}, find the sub-sample {X nm k s 0,x(t ) : k = 1,...,l m } {X n s 0,x(t ) : n = 1,..., N} B rɛ (Y m t,y ) 3. Evaluate Nm=1 ( ) lm k=1 g X nm k s 1,..., X nm k s K,Y mˆt,...,y mˆt K L 1 1 Ym T Xnm k t ɛ Y T m Ĥ ɛ,m,n Nm=1 lm Ym T Xnm k t ɛ Y T m k=1 K Forward-reverse representation for conditional diffusions Page 20 (26)
41 Implementation Assume K has compact support in B r (0) Algorithm 1. Simulate N indep. trajectories (X n ) N n=1 started at X s 0 = x and (Y m ) N m=1 started at Y t = y on D [s 0,t ] and D [t,t], resp. 2. For fixed m {1,..., N}, find the sub-sample {X nm k s 0,x(t ) : k = 1,...,l m } {X n s 0,x(t ) : n = 1,..., N} B rɛ (Y m t,y ) 3. Evaluate Nm=1 ( ) lm k=1 g X nm k s 1,..., X nm k s K,Y mˆt,...,y mˆt K L 1 1 Ym T Xnm k t ɛ Y T m Ĥ ɛ,m,n Nm=1 lm Ym T Xnm k t ɛ Y T m k=1 K Forward-reverse representation for conditional diffusions Page 20 (26)
42 Complexity Assume that X s0,x(s), ( Y t,y(t),y t,y(t) ) can be simulated exactly at constant cost. Cost of simulation step: O(N) Cost of box-ordering step: O(N log(n)) Cost of Evaluation step: O ( N 2 ɛ d) Complexity estimate Case d 4: Choose ɛ = ( N/log N ) 1/d, achieve MSE O(N 1 ) at cost O(N log N) Case d > 4: Choose ɛ = N 2/(4+d), achieve MSE O(N 8/(4+d) ) at cost O(N log N) Forward-reverse representation for conditional diffusions Page 21 (26)
43 Complexity Assume that X s0,x(s), ( Y t,y(t),y t,y(t) ) can be simulated exactly at constant cost. Cost of simulation step: O(N) Cost of box-ordering step: O(N log(n)) Cost of Evaluation step: O ( N 2 ɛ d) Complexity estimate Case d 4: Choose ɛ = ( N/log N ) 1/d, achieve MSE O(N 1 ) at cost O(N log N) Case d > 4: Choose ɛ = N 2/(4+d), achieve MSE O(N 8/(4+d) ) at cost O(N log N) Forward-reverse representation for conditional diffusions Page 21 (26)
44 Complexity Assume that X s0,x(s), ( Y t,y(t),y t,y(t) ) can be simulated exactly at constant cost. Cost of simulation step: O(N) Cost of box-ordering step: O(N log(n)) Cost of Evaluation step: O ( N 2 ɛ d) Complexity estimate Case d 4: Choose ɛ = ( N/log N ) 1/d, achieve MSE O(N 1 ) at cost O(N log N) Case d > 4: Choose ɛ = N 2/(4+d), achieve MSE O(N 8/(4+d) ) at cost O(N log N) Forward-reverse representation for conditional diffusions Page 21 (26)
45 Outline 1 Introduction 2 The forward-reverse method for transition density estimation 3 A forward-reverse representation for conditional expectations 4 The forward-reverse estimator 5 Numerical examples Forward-reverse representation for conditional diffusions Page 22 (26)
46 Conditional expectations of the realized variance Heston model: ds t = µs t dt + v t S t db 1 t, dv t = (αv t + β)dt + ξ v t (ρdb 1 t + ) 1 ρ 2 db 2 t a(x) = µx 1, σ(x) = x 1 x2 0 αx 2 + β ξρ x 2 ξ 1 ρ 2 x 2 Reverse drift: α(x) = (2x 2 + ρξ µ)x 1 (ρξ α)x 2 + ξ 2, c(x) = x 2 + ρξ µ α. β Realized variance: RV 30 i=1 Objective: E[RV S T = s], T = 1/12 ( log(s ti+1 ) log(s ti ) ) 2 Forward-reverse representation for conditional diffusions Page 23 (26)
47 Conditional expectations of the realized variance Heston model: ds t = µs t dt + v t S t db 1 t, dv t = (αv t + β)dt + ξ v t (ρdb 1 t + ) 1 ρ 2 db 2 t a(x) = µx 1, σ(x) = x 1 x2 0 αx 2 + β ξρ x 2 ξ 1 ρ 2 x 2 Reverse drift: α(x) = (2x 2 + ρξ µ)x 1 (ρξ α)x 2 + ξ 2, c(x) = x 2 + ρξ µ α. β Realized variance: RV 30 i=1 Objective: E[RV S T = s], T = 1/12 ( log(s ti+1 ) log(s ti ) ) 2 Forward-reverse representation for conditional diffusions Page 23 (26)
48 Conditional expectations of the realized variance Heston model: ds t = µs t dt + v t S t db 1 t, dv t = (αv t + β)dt + ξ v t (ρdb 1 t + ) 1 ρ 2 db 2 t a(x) = µx 1, σ(x) = x 1 x2 0 αx 2 + β ξρ x 2 ξ 1 ρ 2 x 2 Reverse drift: α(x) = (2x 2 + ρξ µ)x 1 (ρξ α)x 2 + ξ 2, c(x) = x 2 + ρξ µ α. β Realized variance: RV 30 i=1 Objective: E[RV S T = s], T = 1/12 ( log(s ti+1 ) log(s ti ) ) 2 Forward-reverse representation for conditional diffusions Page 23 (26)
49 Numerical experiment Relative MSE 1e 04 1e 02 1e+00 t x = 15 t x = N Forward-reverse representation for conditional diffusions Page 24 (26)
50 Outlook: application to EM algorithms (nutshell sketch) An EM algorithm for parameter estimation of a diffusion can be decomposed into problems of the following structure. ( Consider a diffusion X 0,x (s) ) P 0 s T θ with unknown parameter θ observed at X 0,x (T) = y. Let g(x 1,..., x K,θ) be some suitable and known (likelihood) function. Goal: Compute for a finer grid 0 = s 0,..., s K+1 = T the estimator θ := argmaxe [ θ g(x 0,x (s 1 ),..., X 0,x (s K ),θ) X 0,x (T) = x ] θ via the following iterative algorithm: Initial guess θ 0. After having determined θ n compute [ θ n+1 = argmaxe θ n g(x 0,x (s 1 ),..., X 0,x (s K ),θ) X 0,x (T) = x ] θ Forward-reverse representation for conditional diffusions Page 25 (26)
51 References Bayer, C., Schoenmakers, J.: Simulation of conditional diffusions via forward-reverse stochastic representations, WIAS prepr. 1764, New V. at Bladt, M., Sørensen, M.: Simple simulation of diffusion bridges with application to likelihood inference for diffusions, preprint Delyon, B., Hu, Y.: Simulation of conditioned diffusion and application to parameter estimation, SPA 116(11), , Kusuoka, S., Stroock, D.: Applications of the Malliavin calculus, Part II, J. Fac. Sci. Univ. Yokyo Sec. 1A 32, 1 76, Milstein, G., Schoenmakers, J. G. M., Spokoiny, V.: Transition density estimation for stochastic differential equations via forward reverse representations, Bernoulli 10(2), , Forward-reverse representation for conditional diffusions Page 26 (26)
Stochastic Calculus February 11, / 33
Martingale Transform M n martingale with respect to F n, n =, 1, 2,... σ n F n (σ M) n = n 1 i= σ i(m i+1 M i ) is a Martingale E[(σ M) n F n 1 ] n 1 = E[ σ i (M i+1 M i ) F n 1 ] i= n 2 = σ i (M i+1 M
More informationI forgot to mention last time: in the Ito formula for two standard processes, putting
I forgot to mention last time: in the Ito formula for two standard processes, putting dx t = a t dt + b t db t dy t = α t dt + β t db t, and taking f(x, y = xy, one has f x = y, f y = x, and f xx = f yy
More information1. Stochastic Process
HETERGENEITY IN QUANTITATIVE MACROECONOMICS @ TSE OCTOBER 17, 216 STOCHASTIC CALCULUS BASICS SANG YOON (TIM) LEE Very simple notes (need to add references). It is NOT meant to be a substitute for a real
More informationp 1 ( Y p dp) 1/p ( X p dp) 1 1 p
Doob s inequality Let X(t) be a right continuous submartingale with respect to F(t), t 1 P(sup s t X(s) λ) 1 λ {sup s t X(s) λ} X + (t)dp 2 For 1 < p
More informationGaussian processes for inference in stochastic differential equations
Gaussian processes for inference in stochastic differential equations Manfred Opper, AI group, TU Berlin November 6, 2017 Manfred Opper, AI group, TU Berlin (TU Berlin) inference in SDE November 6, 2017
More informationMathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( )
Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio (2014-2015) Etienne Tanré - Olivier Faugeras INRIA - Team Tosca November 26th, 2014 E. Tanré (INRIA - Team Tosca) Mathematical
More informationWEYL S LEMMA, ONE OF MANY. Daniel W. Stroock
WEYL S LEMMA, ONE OF MANY Daniel W Stroock Abstract This note is a brief, and somewhat biased, account of the evolution of what people working in PDE s call Weyl s Lemma about the regularity of solutions
More informationExact Simulation of Diffusions and Jump Diffusions
Exact Simulation of Diffusions and Jump Diffusions A work by: Prof. Gareth O. Roberts Dr. Alexandros Beskos Dr. Omiros Papaspiliopoulos Dr. Bruno Casella 28 th May, 2008 Content 1 Exact Algorithm Construction
More informationReflected Brownian Motion
Chapter 6 Reflected Brownian Motion Often we encounter Diffusions in regions with boundary. If the process can reach the boundary from the interior in finite time with positive probability we need to decide
More informationTheoretical Tutorial Session 2
1 / 36 Theoretical Tutorial Session 2 Xiaoming Song Department of Mathematics Drexel University July 27, 216 Outline 2 / 36 Itô s formula Martingale representation theorem Stochastic differential equations
More informationStochastic Differential Equations.
Chapter 3 Stochastic Differential Equations. 3.1 Existence and Uniqueness. One of the ways of constructing a Diffusion process is to solve the stochastic differential equation dx(t) = σ(t, x(t)) dβ(t)
More informationIntroduction to numerical simulations for Stochastic ODEs
Introduction to numerical simulations for Stochastic ODEs Xingye Kan Illinois Institute of Technology Department of Applied Mathematics Chicago, IL 60616 August 9, 2010 Outline 1 Preliminaries 2 Numerical
More informationModel Specification Testing in Nonparametric and Semiparametric Time Series Econometrics. Jiti Gao
Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics Jiti Gao Department of Statistics School of Mathematics and Statistics The University of Western Australia Crawley
More informationExercises. T 2T. e ita φ(t)dt.
Exercises. Set #. Construct an example of a sequence of probability measures P n on R which converge weakly to a probability measure P but so that the first moments m,n = xdp n do not converge to m = xdp.
More informationDiscretization of SDEs: Euler Methods and Beyond
Discretization of SDEs: Euler Methods and Beyond 09-26-2006 / PRisMa 2006 Workshop Outline Introduction 1 Introduction Motivation Stochastic Differential Equations 2 The Time Discretization of SDEs Monte-Carlo
More informationHomogenization with stochastic differential equations
Homogenization with stochastic differential equations Scott Hottovy shottovy@math.arizona.edu University of Arizona Program in Applied Mathematics October 12, 2011 Modeling with SDE Use SDE to model system
More informationOn continuous time contract theory
Ecole Polytechnique, France Journée de rentrée du CMAP, 3 octobre, 218 Outline 1 2 Semimartingale measures on the canonical space Random horizon 2nd order backward SDEs (Static) Principal-Agent Problem
More informationBrownian Motion. An Undergraduate Introduction to Financial Mathematics. J. Robert Buchanan. J. Robert Buchanan Brownian Motion
Brownian Motion An Undergraduate Introduction to Financial Mathematics J. Robert Buchanan 2010 Background We have already seen that the limiting behavior of a discrete random walk yields a derivation of
More informationWeak solutions of mean-field stochastic differential equations
Weak solutions of mean-field stochastic differential equations Juan Li School of Mathematics and Statistics, Shandong University (Weihai), Weihai 26429, China. Email: juanli@sdu.edu.cn Based on joint works
More informationSelected Exercises on Expectations and Some Probability Inequalities
Selected Exercises on Expectations and Some Probability Inequalities # If E(X 2 ) = and E X a > 0, then P( X λa) ( λ) 2 a 2 for 0 < λ
More informationHalf of Final Exam Name: Practice Problems October 28, 2014
Math 54. Treibergs Half of Final Exam Name: Practice Problems October 28, 24 Half of the final will be over material since the last midterm exam, such as the practice problems given here. The other half
More informationStrong Markov property of determinantal processes associated with extended kernels
Strong Markov property of determinantal processes associated with extended kernels Hideki Tanemura Chiba university (Chiba, Japan) (November 22, 2013) Hideki Tanemura (Chiba univ.) () Markov process (November
More informationApproximation of BSDEs using least-squares regression and Malliavin weights
Approximation of BSDEs using least-squares regression and Malliavin weights Plamen Turkedjiev (turkedji@math.hu-berlin.de) 3rd July, 2012 Joint work with Prof. Emmanuel Gobet (E cole Polytechnique) Plamen
More informationXt i Xs i N(0, σ 2 (t s)) and they are independent. This implies that the density function of X t X s is a product of normal density functions:
174 BROWNIAN MOTION 8.4. Brownian motion in R d and the heat equation. The heat equation is a partial differential equation. We are going to convert it into a probabilistic equation by reversing time.
More informationStrong Markov property of determinatal processes
Strong Markov property of determinatal processes Hideki Tanemura Chiba university (Chiba, Japan) (August 2, 2013) Hideki Tanemura (Chiba univ.) () Markov process (August 2, 2013) 1 / 27 Introduction The
More informationBridging the Gap between Center and Tail for Multiscale Processes
Bridging the Gap between Center and Tail for Multiscale Processes Matthew R. Morse Department of Mathematics and Statistics Boston University BU-Keio 2016, August 16 Matthew R. Morse (BU) Moderate Deviations
More informationPROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS
PROBABILITY: LIMIT THEOREMS II, SPRING 218. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please
More informationMalliavin Calculus in Finance
Malliavin Calculus in Finance Peter K. Friz 1 Greeks and the logarithmic derivative trick Model an underlying assent by a Markov process with values in R m with dynamics described by the SDE dx t = b(x
More informationA Class of Fractional Stochastic Differential Equations
Vietnam Journal of Mathematics 36:38) 71 79 Vietnam Journal of MATHEMATICS VAST 8 A Class of Fractional Stochastic Differential Equations Nguyen Tien Dung Department of Mathematics, Vietnam National University,
More informationInfinite-dimensional methods for path-dependent equations
Infinite-dimensional methods for path-dependent equations (Università di Pisa) 7th General AMaMeF and Swissquote Conference EPFL, Lausanne, 8 September 215 Based on Flandoli F., Zanco G. - An infinite-dimensional
More informationBrownian Motion. 1 Definition Brownian Motion Wiener measure... 3
Brownian Motion Contents 1 Definition 2 1.1 Brownian Motion................................. 2 1.2 Wiener measure.................................. 3 2 Construction 4 2.1 Gaussian process.................................
More informationMean-field SDE driven by a fractional BM. A related stochastic control problem
Mean-field SDE driven by a fractional BM. A related stochastic control problem Rainer Buckdahn, Université de Bretagne Occidentale, Brest Durham Symposium on Stochastic Analysis, July 1th to July 2th,
More informationOn the Multi-Dimensional Controller and Stopper Games
On the Multi-Dimensional Controller and Stopper Games Joint work with Yu-Jui Huang University of Michigan, Ann Arbor June 7, 2012 Outline Introduction 1 Introduction 2 3 4 5 Consider a zero-sum controller-and-stopper
More informationSome Tools From Stochastic Analysis
W H I T E Some Tools From Stochastic Analysis J. Potthoff Lehrstuhl für Mathematik V Universität Mannheim email: potthoff@math.uni-mannheim.de url: http://ls5.math.uni-mannheim.de To close the file, click
More informationDynamical systems with Gaussian and Levy noise: analytical and stochastic approaches
Dynamical systems with Gaussian and Levy noise: analytical and stochastic approaches Noise is often considered as some disturbing component of the system. In particular physical situations, noise becomes
More informationProperties of an infinite dimensional EDS system : the Muller s ratchet
Properties of an infinite dimensional EDS system : the Muller s ratchet LATP June 5, 2011 A ratchet source : wikipedia Plan 1 Introduction : The model of Haigh 2 3 Hypothesis (Biological) : The population
More informationVast Volatility Matrix Estimation for High Frequency Data
Vast Volatility Matrix Estimation for High Frequency Data Yazhen Wang National Science Foundation Yale Workshop, May 14-17, 2009 Disclaimer: My opinion, not the views of NSF Y. Wang (at NSF) 1 / 36 Outline
More informationRough Burgers-like equations with multiplicative noise
Rough Burgers-like equations with multiplicative noise Martin Hairer Hendrik Weber Mathematics Institute University of Warwick Bielefeld, 3.11.21 Burgers-like equation Aim: Existence/Uniqueness for du
More informationErrata for Stochastic Calculus for Finance II Continuous-Time Models September 2006
1 Errata for Stochastic Calculus for Finance II Continuous-Time Models September 6 Page 6, lines 1, 3 and 7 from bottom. eplace A n,m by S n,m. Page 1, line 1. After Borel measurable. insert the sentence
More informationLAN property for ergodic jump-diffusion processes with discrete observations
LAN property for ergodic jump-diffusion processes with discrete observations Eulalia Nualart (Universitat Pompeu Fabra, Barcelona) joint work with Arturo Kohatsu-Higa (Ritsumeikan University, Japan) &
More informationStochastic continuity equation and related processes
Stochastic continuity equation and related processes Gabriele Bassi c Armando Bazzani a Helmut Mais b Giorgio Turchetti a a Dept. of Physics Univ. of Bologna, INFN Sezione di Bologna, ITALY b DESY, Hamburg,
More informationLecture 4: Ito s Stochastic Calculus and SDE. Seung Yeal Ha Dept of Mathematical Sciences Seoul National University
Lecture 4: Ito s Stochastic Calculus and SDE Seung Yeal Ha Dept of Mathematical Sciences Seoul National University 1 Preliminaries What is Calculus? Integral, Differentiation. Differentiation 2 Integral
More informationNonlinear representation, backward SDEs, and application to the Principal-Agent problem
Nonlinear representation, backward SDEs, and application to the Principal-Agent problem Ecole Polytechnique, France April 4, 218 Outline The Principal-Agent problem Formulation 1 The Principal-Agent problem
More informationThe concentration of a drug in blood. Exponential decay. Different realizations. Exponential decay with noise. dc(t) dt.
The concentration of a drug in blood Exponential decay C12 concentration 2 4 6 8 1 C12 concentration 2 4 6 8 1 dc(t) dt = µc(t) C(t) = C()e µt 2 4 6 8 1 12 time in minutes 2 4 6 8 1 12 time in minutes
More informationStochastic optimal control with rough paths
Stochastic optimal control with rough paths Paul Gassiat TU Berlin Stochastic processes and their statistics in Finance, Okinawa, October 28, 2013 Joint work with Joscha Diehl and Peter Friz Introduction
More informationMan Kyu Im*, Un Cig Ji **, and Jae Hee Kim ***
JOURNAL OF THE CHUNGCHEONG MATHEMATICAL SOCIETY Volume 19, No. 4, December 26 GIRSANOV THEOREM FOR GAUSSIAN PROCESS WITH INDEPENDENT INCREMENTS Man Kyu Im*, Un Cig Ji **, and Jae Hee Kim *** Abstract.
More informationNonparametric Drift Estimation for Stochastic Differential Equations
Nonparametric Drift Estimation for Stochastic Differential Equations Gareth Roberts 1 Department of Statistics University of Warwick Brazilian Bayesian meeting, March 2010 Joint work with O. Papaspiliopoulos,
More informationStochastic Differential Equations
CHAPTER 1 Stochastic Differential Equations Consider a stochastic process X t satisfying dx t = bt, X t,w t dt + σt, X t,w t dw t. 1.1 Question. 1 Can we obtain the existence and uniqueness theorem for
More informationMA8109 Stochastic Processes in Systems Theory Autumn 2013
Norwegian University of Science and Technology Department of Mathematical Sciences MA819 Stochastic Processes in Systems Theory Autumn 213 1 MA819 Exam 23, problem 3b This is a linear equation of the form
More informationOn the Goodness-of-Fit Tests for Some Continuous Time Processes
On the Goodness-of-Fit Tests for Some Continuous Time Processes Sergueï Dachian and Yury A. Kutoyants Laboratoire de Mathématiques, Université Blaise Pascal Laboratoire de Statistique et Processus, Université
More informationProf. Erhan Bayraktar (University of Michigan)
September 17, 2012 KAP 414 2:15 PM- 3:15 PM Prof. (University of Michigan) Abstract: We consider a zero-sum stochastic differential controller-and-stopper game in which the state process is a controlled
More informationMaximum Likelihood Drift Estimation for Gaussian Process with Stationary Increments
Austrian Journal of Statistics April 27, Volume 46, 67 78. AJS http://www.ajs.or.at/ doi:.773/ajs.v46i3-4.672 Maximum Likelihood Drift Estimation for Gaussian Process with Stationary Increments Yuliya
More informationLAN property for sde s with additive fractional noise and continuous time observation
LAN property for sde s with additive fractional noise and continuous time observation Eulalia Nualart (Universitat Pompeu Fabra, Barcelona) joint work with Samy Tindel (Purdue University) Vlad s 6th birthday,
More informationStability of Stochastic Differential Equations
Lyapunov stability theory for ODEs s Stability of Stochastic Differential Equations Part 1: Introduction Department of Mathematics and Statistics University of Strathclyde Glasgow, G1 1XH December 2010
More informationRegularity of the density for the stochastic heat equation
Regularity of the density for the stochastic heat equation Carl Mueller 1 Department of Mathematics University of Rochester Rochester, NY 15627 USA email: cmlr@math.rochester.edu David Nualart 2 Department
More informationA Semi-Lagrangian scheme for a regularized version of the Hughes model for pedestrian flow
A Semi-Lagrangian scheme for a regularized version of the Hughes model for pedestrian flow Adriano FESTA Johann Radon Institute for Computational and Applied Mathematics (RICAM) Austrian Academy of Sciences
More informationCOMPUTER-AIDED MODELING AND SIMULATION OF ELECTRICAL CIRCUITS WITH α-stable NOISE
APPLICATIONES MATHEMATICAE 23,1(1995), pp. 83 93 A. WERON(Wroc law) COMPUTER-AIDED MODELING AND SIMULATION OF ELECTRICAL CIRCUITS WITH α-stable NOISE Abstract. The aim of this paper is to demonstrate how
More informationLecture 4: Numerical Solution of SDEs, Itô Taylor Series, Gaussian Process Approximations
Lecture 4: Numerical Solution of SDEs, Itô Taylor Series, Gaussian Process Approximations Simo Särkkä Aalto University Tampere University of Technology Lappeenranta University of Technology Finland November
More informationLecture 12: Diffusion Processes and Stochastic Differential Equations
Lecture 12: Diffusion Processes and Stochastic Differential Equations 1. Diffusion Processes 1.1 Definition of a diffusion process 1.2 Examples 2. Stochastic Differential Equations SDE) 2.1 Stochastic
More informationOn density asymptotics for diffusions, old and new
On density asymptotics for diffusions, old and new Stefano De Marco (Ecole Polytechnique) Paris, January 2014 S. De Marco (Ecole Polytechnique) 7/01/14 0 / 15 d dx = b(x)dt + σ j (X)dW j, X 0 = x 0, t
More information1 Brownian Local Time
1 Brownian Local Time We first begin by defining the space and variables for Brownian local time. Let W t be a standard 1-D Wiener process. We know that for the set, {t : W t = } P (µ{t : W t = } = ) =
More informationStochastic Differential Equations
Chapter 5 Stochastic Differential Equations We would like to introduce stochastic ODE s without going first through the machinery of stochastic integrals. 5.1 Itô Integrals and Itô Differential Equations
More informationON THE POLICY IMPROVEMENT ALGORITHM IN CONTINUOUS TIME
ON THE POLICY IMPROVEMENT ALGORITHM IN CONTINUOUS TIME SAUL D. JACKA AND ALEKSANDAR MIJATOVIĆ Abstract. We develop a general approach to the Policy Improvement Algorithm (PIA) for stochastic control problems
More information1. Stochastic Processes and filtrations
1. Stochastic Processes and 1. Stoch. pr., A stochastic process (X t ) t T is a collection of random variables on (Ω, F) with values in a measurable space (S, S), i.e., for all t, In our case X t : Ω S
More informationKolmogorov Equations and Markov Processes
Kolmogorov Equations and Markov Processes May 3, 013 1 Transition measures and functions Consider a stochastic process {X(t)} t 0 whose state space is a product of intervals contained in R n. We define
More informationMatemáticas II: Segundo del Grado en Ingeniería Aeroespacial
Matemáticas II: Segundo del Grado en Ingeniería Aeroespacial Sergio Blanes, Dolors Roselló Ecuaciones diferenciales y transformadas de Laplace con aplicaciones L.M. Sánchez, M.P. Legua Ref.: 211-798 Capítulo
More informationA Probability Review
A Probability Review Outline: A probability review Shorthand notation: RV stands for random variable EE 527, Detection and Estimation Theory, # 0b 1 A Probability Review Reading: Go over handouts 2 5 in
More informationPUTNAM PROBLEMS DIFFERENTIAL EQUATIONS. First Order Equations. p(x)dx)) = q(x) exp(
PUTNAM PROBLEMS DIFFERENTIAL EQUATIONS First Order Equations 1. Linear y + p(x)y = q(x) Muliply through by the integrating factor exp( p(x)) to obtain (y exp( p(x))) = q(x) exp( p(x)). 2. Separation of
More informationStochastic Modelling in Climate Science
Stochastic Modelling in Climate Science David Kelly Mathematics Department UNC Chapel Hill dtbkelly@gmail.com November 16, 2013 David Kelly (UNC) Stochastic Climate November 16, 2013 1 / 36 Why use stochastic
More information(1 + 2y)y = x. ( x. The right-hand side is a standard integral, so in the end we have the implicit solution. y(x) + y 2 (x) = x2 2 +C.
Midterm 1 33B-1 015 October 1 Find the exact solution of the initial value problem. Indicate the interval of existence. y = x, y( 1) = 0. 1 + y Solution. We observe that the equation is separable, and
More informationOn Stopping Times and Impulse Control with Constraint
On Stopping Times and Impulse Control with Constraint Jose Luis Menaldi Based on joint papers with M. Robin (216, 217) Department of Mathematics Wayne State University Detroit, Michigan 4822, USA (e-mail:
More informationGeneralized Gaussian Bridges of Prediction-Invertible Processes
Generalized Gaussian Bridges of Prediction-Invertible Processes Tommi Sottinen 1 and Adil Yazigi University of Vaasa, Finland Modern Stochastics: Theory and Applications III September 1, 212, Kyiv, Ukraine
More informationReaction-Diffusion Equations In Narrow Tubes and Wave Front P
Outlines Reaction-Diffusion Equations In Narrow Tubes and Wave Front Propagation University of Maryland, College Park USA Outline of Part I Outlines Real Life Examples Description of the Problem and Main
More informationOptimal Control of Partially Observable Piecewise Deterministic Markov Processes
Optimal Control of Partially Observable Piecewise Deterministic Markov Processes Nicole Bäuerle based on a joint work with D. Lange Wien, April 2018 Outline Motivation PDMP with Partial Observation Controlled
More informationChain Rule. MATH 311, Calculus III. J. Robert Buchanan. Spring Department of Mathematics
3.33pt Chain Rule MATH 311, Calculus III J. Robert Buchanan Department of Mathematics Spring 2019 Single Variable Chain Rule Suppose y = g(x) and z = f (y) then dz dx = d (f (g(x))) dx = f (g(x))g (x)
More informationOptimal Trade Execution with Instantaneous Price Impact and Stochastic Resilience
Optimal Trade Execution with Instantaneous Price Impact and Stochastic Resilience Ulrich Horst 1 Humboldt-Universität zu Berlin Department of Mathematics and School of Business and Economics Vienna, Nov.
More informationContinuous dependence estimates for the ergodic problem with an application to homogenization
Continuous dependence estimates for the ergodic problem with an application to homogenization Claudio Marchi Bayreuth, September 12 th, 2013 C. Marchi (Università di Padova) Continuous dependence Bayreuth,
More informationThe multidimensional Ito Integral and the multidimensional Ito Formula. Eric Mu ller June 1, 2015 Seminar on Stochastic Geometry and its applications
The multidimensional Ito Integral and the multidimensional Ito Formula Eric Mu ller June 1, 215 Seminar on Stochastic Geometry and its applications page 2 Seminar on Stochastic Geometry and its applications
More informationCalculus II. George Voutsadakis 1. LSSU Math 152. Lake Superior State University. 1 Mathematics and Computer Science
Calculus II George Voutsadakis 1 1 Mathematics and Computer Science Lake Superior State University LSSU Math 152 George Voutsadakis (LSSU) Calculus II February 2015 1 / 28 Outline 1 Introduction to Differential
More informationMATH 6605: SUMMARY LECTURE NOTES
MATH 6605: SUMMARY LECTURE NOTES These notes summarize the lectures on weak convergence of stochastic processes. If you see any typos, please let me know. 1. Construction of Stochastic rocesses A stochastic
More informationIntroduction to the Numerical Solution of SDEs Selected topics
0 / 23 Introduction to the Numerical Solution of SDEs Selected topics Andreas Rößler Summer School on Numerical Methods for Stochastic Differential Equations at Vienna University of Technology, Austria
More informationIntroduction to multiscale modeling and simulation. Explicit methods for ODEs : forward Euler. y n+1 = y n + tf(y n ) dy dt = f(y), y(0) = y 0
Introduction to multiscale modeling and simulation Lecture 5 Numerical methods for ODEs, SDEs and PDEs The need for multiscale methods Two generic frameworks for multiscale computation Explicit methods
More informationA new approach for investment performance measurement. 3rd WCMF, Santa Barbara November 2009
A new approach for investment performance measurement 3rd WCMF, Santa Barbara November 2009 Thaleia Zariphopoulou University of Oxford, Oxford-Man Institute and The University of Texas at Austin 1 Performance
More informationApplications of Ito s Formula
CHAPTER 4 Applications of Ito s Formula In this chapter, we discuss several basic theorems in stochastic analysis. Their proofs are good examples of applications of Itô s formula. 1. Lévy s martingale
More informationExercises for Multivariable Differential Calculus XM521
This document lists all the exercises for XM521. The Type I (True/False) exercises will be given, and should be answered, online immediately following each lecture. The Type III exercises are to be done
More informationSDE Coefficients. March 4, 2008
SDE Coefficients March 4, 2008 The following is a summary of GARD sections 3.3 and 6., mainly as an overview of the two main approaches to creating a SDE model. Stochastic Differential Equations (SDE)
More informationUniformly Uniformly-ergodic Markov chains and BSDEs
Uniformly Uniformly-ergodic Markov chains and BSDEs Samuel N. Cohen Mathematical Institute, University of Oxford (Based on joint work with Ying Hu, Robert Elliott, Lukas Szpruch) Centre Henri Lebesgue,
More informationLecture No 2 Degenerate Diffusion Free boundary problems
Lecture No 2 Degenerate Diffusion Free boundary problems Columbia University IAS summer program June, 2009 Outline We will discuss non-linear parabolic equations of slow diffusion. Our model is the porous
More informationBrownian Motion: Fokker-Planck Equation
Chapter 7 Brownian Motion: Fokker-Planck Equation The Fokker-Planck equation is the equation governing the time evolution of the probability density of the Brownian particla. It is a second order differential
More informationENGIN 211, Engineering Math. Laplace Transforms
ENGIN 211, Engineering Math Laplace Transforms 1 Why Laplace Transform? Laplace transform converts a function in the time domain to its frequency domain. It is a powerful, systematic method in solving
More informationUCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, Practice Final Examination (Winter 2017)
UCSD ECE250 Handout #27 Prof. Young-Han Kim Friday, June 8, 208 Practice Final Examination (Winter 207) There are 6 problems, each problem with multiple parts. Your answer should be as clear and readable
More informationSmall-time asymptotics of stopped Lévy bridges and simulation schemes with controlled bias
Small-time asymptotics of stopped Lévy bridges and simulation schemes with controlled bias José E. Figueroa-López 1 1 Department of Statistics Purdue University Seoul National University & Ajou University
More informationExistence, Uniqueness and Stability of Invariant Distributions in Continuous-Time Stochastic Models
Existence, Uniqueness and Stability of Invariant Distributions in Continuous-Time Stochastic Models Christian Bayer and Klaus Wälde Weierstrass Institute for Applied Analysis and Stochastics and University
More informationSolvability of G-SDE with Integral-Lipschitz Coefficients
Solvability of G-SDE with Integral-Lipschitz Coefficients Yiqing LIN joint work with Xuepeng Bai IRMAR, Université de Rennes 1, FRANCE http://perso.univ-rennes1.fr/yiqing.lin/ ITN Marie Curie Workshop
More informationJoint Parameter Estimation of the Ornstein-Uhlenbeck SDE driven by Fractional Brownian Motion
Joint Parameter Estimation of the Ornstein-Uhlenbeck SDE driven by Fractional Brownian Motion Luis Barboza October 23, 2012 Department of Statistics, Purdue University () Probability Seminar 1 / 59 Introduction
More informationJump-type Levy Processes
Jump-type Levy Processes Ernst Eberlein Handbook of Financial Time Series Outline Table of contents Probabilistic Structure of Levy Processes Levy process Levy-Ito decomposition Jump part Probabilistic
More informationReal Time Stochastic Control and Decision Making: From theory to algorithms and applications
Real Time Stochastic Control and Decision Making: From theory to algorithms and applications Evangelos A. Theodorou Autonomous Control and Decision Systems Lab Challenges in control Uncertainty Stochastic
More informationMATH 4B Differential Equations, Fall 2016 Final Exam Study Guide
MATH 4B Differential Equations, Fall 2016 Final Exam Study Guide GENERAL INFORMATION AND FINAL EXAM RULES The exam will have a duration of 3 hours. No extra time will be given. Failing to submit your solutions
More informationA Short Introduction to Diffusion Processes and Ito Calculus
A Short Introduction to Diffusion Processes and Ito Calculus Cédric Archambeau University College, London Center for Computational Statistics and Machine Learning c.archambeau@cs.ucl.ac.uk January 24,
More informationn! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2
Order statistics Ex. 4. (*. Let independent variables X,..., X n have U(0, distribution. Show that for every x (0,, we have P ( X ( < x and P ( X (n > x as n. Ex. 4.2 (**. By using induction or otherwise,
More information