DOOB S DECOMPOSITION THEOREM FOR NEAR-SUBMARTINGALES

Size: px
Start display at page:

Download "DOOB S DECOMPOSITION THEOREM FOR NEAR-SUBMARTINGALES"

Transcription

1 Communications on Stochastic Analysis Vol. 9, No. 4 (215) Serials Publications DOOB S DECOMPOSITION THEOREM FOR NEAR-SUBMARTINGALES HUI-HSIUNG KUO AND KIMIAKI SAITÔ* Abstract. We study the discrete parameter case of near-martingales, nearsubmartingales, and near-supermartingales. In particular, we prove Doob s decomposition theorem for near-submartingales. This generalizes the classical case for submartingales. 1. Motivation From Non-adapted Stochastic Integral Let B(t), t, be a Brownian motion starting at and {F t the filtration given by B(t), namely, F t = σ{b(s); s t, t. The Itô integral b f(t) db(t) a (see, e.g., the book [8]) is defined for {F t -adapted stochastic processes f(t) with almost all sample paths being in L 2 [a, b]. Several extensions of the Itô theory of stochastic integration to cover non-adapted integrands have been introduced and extensively studied by, just to mention a few names, Buckdahn [3], Dorogovtsev [4], Hitsuda [5], Itô [6], Kuo Potthoff [1], León Protter [12], Nualart Pardoux [13], Pardoux Protter [14], Russo Vallois [15], and Skorokhod [16]. In particular, in his lecture for the 1976 Kyoto Symposium, Itô [6] gave rather elegant ideas to define the following non-adaptive stochastic integral (I) B(1) db(s), t 1, (1.1) namely, enlarging the σ-field F t to G t = σ{b(1), B(s); s t, t 1, so that the integrand B(1) is adaptive and B(t) is a quasimartingale with respect to the filtration {G t. Then the stochastic integral in equation (1.1) is defined as a stochastic integral with respect to a quasimartingale and has the value (I) B(1) db(s) = B(1)B(t), t 1. (1.2) On the other hand, the Hitsuda Skorokhod integral (see [5] [16]) can be expressed in terms of a white noise integral (see the book [7]) and has the value (HS) B(1) db(s) = s B(1) ds = B(1)B(t) t, t 1. (1.3) Received ; Communicated by the editors. 21 Mathematics Subject Classification. Primary 6G42, 6G48; Secondary 6G5, 6H5. Key words and phrases. Brownian motion, stochastic integral, Hitsuda Skorokhod integral, conditional expectation, martingale, near-martingale, near-submartingale, near-supermartingale, Doob s decomposition theorem, instantly independent sequence. *This work was supported by JSPS Grant-in-Aid Scientific Research 15K

2 468 HUI-HSIUNG KUO AND KIMIAKI SAITÔ Being motivated by Itô s ideas and observing the different values of equations (1.2) and (1.3), we have defined in [1] [2] the stochastic integral B(1) db(s) in the following way. Decompose the integrand B(1) as B(1) = B(t) + ( B(1) B(t) ), where the first term B(t) is the Itô part of B(1) and the second term B(1) B(t) is the counterpart of B(1). For the Itô part, the evaluation points are the left endpoints of subintervals, while the evaluation points for the counterpart are the right endpoints of subintervals. Thus for t 1, we have [ ( )] B(1) db(s) = B(s) + B(1) B(s) db(s) = lim n i=1 = lim n i=1 = lim n n [ B(si 1 ) + ( B(1) B(s i ) )]( B(s i ) B(s i 1 ) ) n [ ( B(1) B(si ) B(s i 1 ) )]( B(s i ) B(s i 1 ) ) ( B(1) n ( B(si ) B(s i 1 ) ) i=1 n ( B(si ) B(s i 1 ) ) ) 2 = B(1)B(t) t, (1.4) where the limit is convergence in probability. Note that this value is the same as the Hitsuda Skorokhod integral in equation (1.3). There is an intrinsic difference between the stochastic processes X t = B(1)B(t) t, Y t = B(1)B(t), t 1, (1.5) given by equations (1.4) and (1.2), respectively. For any s t, we see that In particular, put t = s to get It follows from equations (1.6) and (1.7) that i=1 E[X t F s ] = B(s) 2 s. (1.6) E[X s F s ] = B(s) 2 s. (1.7) E[X t F s ] = E[X s F s ], s t. (1.8) On the other hand, it is easy to check that the stochastic process Y t = B(1)B(t) in equation (1.5) does not satisfy equation (1.8). This leads to the following concept introduced in [11]. Definition 1.1. A stochastic process X t with E X t < for a t b is called a near-martingale with respect to a filtration {F t if it satisfies the condition in equation (1.8). We can define near-submartingale and near-supermartingale with respect to a filtration {F t by the following respective conditions: and E[X t F s ] E[X s F s ], s t, (1.9) E[X t F s ] E[X s F s ], s t.

3 DOOB S DECOMPOSITION THEOREM FOR NEAR-SUBMARTINGALES 469 Observe that if a stochastic process X t is adapted to a filtration {F t, then nearmartingale, near-submartingale, and near-supermartingale reduce to martingale, submartingale, and supermartingale, respectively. In this paper we will study the discrete parameter case of near-martingales and near-submartingales. In particular, we will prove Doob s decomposition theorem for near-submartingales. 2. Near-martingales and Near-submartingales Let {F n ; 1 n N be a fixed filtration, i.e., an increasing sequence of σ-fields. Definition 2.1. A sequence X n, 1 n N, of integrable random variables is called a near-martingale with respect to {F n ; 1 n N if E[X n+1 F n ] = E[X n F n ], 1 n N 1. (2.1) Remark 2.2. It is easy to see that the equality in equation (2.1) is equivalent to the equality: E[X m F n ] = E[X n F n ], 1 n m N. (2.2) Similarly, we can define near-submartingale and near-supermartingale just by replacing the equality sign in equation (2.1) with and, respectively. They also have the corresponding equivalent conditions as in equation (2.2). Obviously, if a sequence X n, 1 n N, is adapted to {F n ; 1 n N, then near-martingale, near-submartingale, and near-supermartingale are martingale, submartingale, and supermartingale, respectively. Example 2.3. Take a sequence ξ 1, ξ 2,..., ξ N of independent random variables with mean. Let {F n be the filtration given by F n = σ{ξ k ; 1 k n. Put S n = ξ ξ n, X n = S N S n, 1 n N. (2.3) The sequence S n, 1 n N, is a martingale. On the other hand, E[X n+1 F n ] = E[ξ n ξ N F n ] = E(ξ n ξ N ) =. Similarly, we have E[X n F n ] =. Thus E[X n+1 F n ] = E[X n F n ], which shows that X n, 1 n N, is a near-martingale. Furthermore, suppose ξ n, n 1, is a sequence of independent random variables with mean. For fixed N, X n = S N S n, 1 n N, is a near-martingale as shown above. However, X n = S N S n, n N, is a martingale. Example 2.4. Let ξ 1, ξ 2,..., ξ N be a sequence of independent random variables with mean and var(ξ n ) = σ 2 n. Let F n = σ{ξ k ; 1 k n. Put S n = ξ ξ n, X n = S n S N n σk, 2 1 n N. (2.4)

4 47 HUI-HSIUNG KUO AND KIMIAKI SAITÔ It is easy to check that [ n+1 E[X n+1 F n ] = E S n+1 S N σ 2 k ] F n = E[(S n + ξ n+1 )(S n + ξ n+1 + ξ N ) n+1 F n ] n+1 = Sn 2 + σn+1 2 = S 2 n Similarly, we can easily derive σ 2 k n σk. 2 (2.5) E[X n F n ] = S 2 n σ 2 k n σk. 2 (2.6) It follows from equations (2.5) and (2.6) that E[X n+1 F n ] = E[X n F n ]. Hence the sequence X n = S n S N n σ2 k, 1 n N, is a near-martingale. Moreover, let ξ n, n 1, be a sequence of independent random variables with mean and var(ξ n ) = σn. 2 Take F n = σ{ξ k ; 1 k n. Define S n and X n as in equation (2.4). For fixed N, the sequence X n, 1 n N, is a near-martingale as shown above. On the other hand, the sequence X n, n N, is a martingale. Theorem 2.5. Let S n, 1 n N, be a square integrable martingale with respect to a filtration {F n ; 1 n N. Then is a near-martingale. Proof. Note that Hence we have V n = S n (S N S n ), 1 n N, V n+1 V n = (S n+1 S n )S N S 2 n+1 + S 2 n. (2.7) E[V n+1 V n F n ] = E[(S n+1 S n )S N F n ] E[S 2 n+1 F n ] + E[S 2 n F n ] = E { E[(S n+1 S n )S N F n+1 ] F n E[S 2 n+1 F n ] + S 2 n = E { (S n+1 S n )E[S N F n+1 ] F n E[S 2 n+1 F n ] + S 2 n = E { (S n+1 S n )S n+1 F n E[S 2 n+1 F n ] + S 2 n = S n E[S n+1 F n ] + S 2 n = S 2 n + S 2 n =. Hence E[V n+1 F n ] = E[V n F n ] and so V n, 1 n N, is a near-martingale.

5 DOOB S DECOMPOSITION THEOREM FOR NEAR-SUBMARTINGALES 471 Theorem 2.6. Suppose S n, n = 1, 2,..., is a square integrable martingale with respect to a filtration {F n ; 1 n N. For a fixed natural number N, let V n = S n (S N S n ), n = 1, 2,.... Then (1) V n, 1 n N, is a near-martingale, (2) V n, n N, is a supermartingale. Proof. The first assertion follows from Theorem 2.5. To prove the second assertion, we use equation (2.7) to show that for n N, E[V n+1 V n F n ] = S N E[(S n+1 S n ) F n ] E[S 2 n+1 F n ] + S 2 n = E[S 2 n+1 F n ] + S 2 n, since S 2 n is a submartingale. Thus E[V n+1 F n ] E[V n F n ] for n N. But the sequence V n, n N, is adapted to the filtration {F n. Therefore, we have E[V n+1 F n ] V n, n N. This shows that V n, n N, is a supermartingale. 3. Doob s Decomposition Theorem In this section we prove Doob s decomposition theorem for near-submartingales. Theorem 3.1. Let X n, n 1, be a near-submartingale with respect to a filtration {F n. Then there exists a unique decomposition with M n and A n satisfying the following conditions: (1) M n, n 1, is a near-martingale. (2) A 1 =. (3) A n is F n 1 -measurable for n 2. (4) A n is inceasing almost surely. X n = M n + A n, n 1, (3.1) Proof. Existence of a decomposition Define A 1 = and M 1 = X 1. Then we have equation (3.1) for n = 1. To find A 2 and M 2 such that X 2 = M 2 + A 2 with desired properties, we take conditional expectation with respect to F 1 : Therefore, we define E[X 2 F 1 ] = E[M 2 F 1 ] + E[A 2 F 1 ] = E[M 1 F 1 ] + A 2 = E[X 1 F 1 ] + A 2. A 2 = E[X 2 F 1 ] E[X 1 F 1 ], M 2 = X 2 A 2. Then we have equation (3.1) for n = 2. Observe that A 2 is F 1 -measurable and A 1 A 2 almost surely since {X n is a near-submartingale.

6 472 HUI-HSIUNG KUO AND KIMIAKI SAITÔ Inductively, we repeat the above arguments to define A n and M n for n 3 by n ( ) A n = E[X k F k 1 ] E[X k 1 F k 1 ], k=2 M n = X n A n. Then we have equation (3.1) for n 3. Notice that A n is F n 1 -measurable and A n 1 A n almost surely since {X n is a near-submartingale. Now, we need to show that M n, n 1, is a near-martingale with respect to {F n. Note that for n 2, we have M n = X n which yields the equality n k=2 ( ) E[X k F k 1 ] E[X k 1 F k 1 ], M n M n 1 = X n X n 1 E[X n F n 1 ] + E[X n 1 F n 1 ]. Then we take conditional expectation with respect to F n 1 to show that E[M n M n 1 F n 1 ] =, namely, E[M n F n 1 ] = E[M n 1 F n 1 ]. Hence M n, n 1, is a near-martingale with respect to {F n. Uniqueness of a decomposition Suppose we have two such decompositions Then we have X n = M n + A n = N n + B n, n 1. (3.2) M n N n = B n A n, n 1. (3.3) For n = 1, we have B 1 = A 1 =. Hence M 1 = N 1. For n 2, take the conditional expectation of equation (3.3) with respect to F n 1 to get E[M n N n F n 1 ] = E[B n A n F n 1 ] = B n A n, (3.4) where in the last equality we have used the fact that A n and B n are F n 1 - measurable. On the other hand, use equation (3.3) for n 1 and the fact that M n and N n are near-martingales to get E[M n N n F n 1 ] = E[M n 1 N n 1 F n 1 ] = E[B n 1 A n 1 F n 1 ] = B n 1 A n 1, (3.5) where the last equality holds since B n 1 and A n 1 are F n 2 -measurable and so are F n 1 -measurable. Thus by equations (3.4) and (3.5), B n A n = B n 1 A n 1, n 2. This equation together with A 1 = B 1 implies that A n = B n almost surely for all n 1. Then by equation (3.2) we have M n = N n almost surely for all n 1. Hence the decomposition is unique.

7 DOOB S DECOMPOSITION THEOREM FOR NEAR-SUBMARTINGALES 473 Example 3.2. Let ξ n, n 1, be a sequence of independent random variables with mean and var(ξ n ) = σ 2 n. Take F n = σ{ξ k ; 1 k n. Define S n = ξ ξ n. For fixed N, consider the sequence X n = S n S N, 1 n N. (3.6) First we show that the sequence X n, 1 n N, is a near-submartingale. It is easy to see that On the other hand, we have E[X n+1 F n ] = E[S n+1 S N F n ] = E[(S n + ξ n+1 )(ξ ξ N ) F n ] = E[(S n + ξ n+1 ) 2 F n ] = E[S 2 n + 2S n ξ n+1 + ξ 2 n+1 F n ] = S 2 n + σ 2 n+1. (3.7) E[X n F n ] = E[S n S N F n ] = S n E[S N F n ] = S 2 n. (3.8) By equations (3.7) and (3.8), we have E[X n+1 F n ] E[X n F n ] almost surely. Hence X n, 1 n N, is a near-submartingale. To find the Doob decomposition of X n, 1 n N, recall from Example 2.4 that the sequence n Z n S n S N σk, 2 1 n N, is a near-martingale. This motivates us to define M n and A n by and { S1 S N, if n = 1, M n = S n S N n k=2 σ2 k, if n 2. A n = {, if n = 1, n k=2 σ2 k, if n 2. Note that M n = Z n + σ 2 1. Hence M n is a near-martingale. Then we can easily see that the Doob decomposition of S n S N is given by S n S N = M n + A n, 1 n N. We need to point out a difference between martingale case and near-martingale case. Suppose X n is a square integrable martingale. It is well known that Xn 2 is a submartingale. However, for a square integrable near-martingale X n, it is not true in general that Xn 2 is a near-submartingale. For instance, the sequence X n = S N S n, 1 n N, in Example 2.3 is a near-martingale. However, it is easy to check that Xn, 2 1 n N, is not a near-submartingale. In fact, it is a near-supermartingale.

8 474 HUI-HSIUNG KUO AND KIMIAKI SAITÔ 4. Instantly Independent Sequences Note that martingales must be adapted with respect to an associated filtration. In [11], we introduced the concept of instantly independent stochastic processes, which play the counterpart role of adapted stochastic processes. Thus for the discrete case, we have instantly independent sequences of random variables. Definition 4.1. A sequence {Φ n of random variables is said to be instantly independent with respect to a filtration {F n if Φ n and F n are independent for each n. We have the following two basic properties of instantly independent sequences of random variables. Theorem 4.2. If X n is a near-martingale, then EX n is a constant (independent of n). Conversely, if EX n is a constant and X n is instantly independent, then X n is a near-martingale. Proof. Suppose X n is a near-martingale. Then we have E[X n+1 F n ] = E[X n F n ], n 1. Upon taking expectation, we immediately get EX n+1 = EX n for all n 1. Hence EX n is a constant. Conversely, suppose EX n is a constant and X n is instantly independent with respect to a filtration {F n. Then E[X n+1 F n ] = E { E[X n+1 F n+1 ] F n = E { EX n+1 F n = EX n+1 = c, where c is a constant. On the other hand, since X n and F n are independent, we have E[X n F n ] = EX n = c. Hence E[X n+1 F n ] = E[X n F n ] and so X n, n 1, is a near-martingale. Theorem 4.3. Suppose X n is a square integrable martingale and Φ n is a square integrable sequence of instantly independent random variables with EΦ n being a constant. Then the product X n Φ n is a near-martingale. Proof. Using the assumptions we can easily derive E[X n+1 Φ n+1 F n ] = E { E[X n+1 Φ n+1 F n+1 ] F n = E { X n+1 E[Φ n+1 F n+1 ] Fn = E { X n+1 EΦ n+1 F n = EΦ n+1 E[X n+1 F n ] = cx n, (4.1) where c = EΦ n is a constant. On the other hand, we have E[X n Φ n F n ] = X n E[Φ n F n ] = X n EΦ n = cx n. (4.2)

9 DOOB S DECOMPOSITION THEOREM FOR NEAR-SUBMARTINGALES 475 It follows from equations (4.1) and (4.2) that E[X n+1 Φ n+1 F n ] = E[X n Φ n F n ] almost surely. Hence X n is a near-martingale. Example 4.4. Let ξ 1, ξ 2,..., ξ N be a sequence of independent random variables with mean and finite variances. Let F n = σ{ξ k ; 1 k n. Put S n = ξ 1 + ξ ξ n. Then S n is a martingale with respect to the filtration {F n. Let θ be a real-vlaued function on R. For fixed N, assume that the random variables θ(s N S n ), 1 n N, are square integrable. Then the following sequence Φ n = θ(s N S n ) Eθ(S N S n ), 1 n N, is instantly independent with respect to the filtration {F n with mean. Hence by Theorem 4.3 the sequence ) Y n = S n (θ(s N S n ) Eθ(S N S n ), 1 n N, is a near-martingale. Acknowledgment. The mathematical concepts and the results in this paper were obtained in many discussions with K. Saitô during Kuo s visits to Meijo University since 211. Kuo would like to give his deepest appreciation to Professor Saitô for the invitations and for the warm hospitality. References 1. Ayed, W. and Kuo, H.-H.: An extension of the Itô integral, Communications on Stochastic Analysis 2, no. 3 (28) Ayed, W. and Kuo, H.-H.: An extension of the Itô integral: toward a general theory of stochastic integration, Theory of Stochastic Processes 16(32), no. 1 (21) Buckdahn, R.: Anticipative Girsanov transformations, Probab. Th. Rel. Fields 89 (1991) Dorogovtsev, A. A.: Itô Volterra equations with an anticipating right-hand side in the absence of moments, Infinite-dimensional Stochastic Analysis (Russian) 41 5, Akad. Nauk Ukrain. SSR, Inst. Mat., Kiev, Hitsuda, M.: Formula for Brownian partial derivatives, Second Japan-USSR Symp. Probab. Th.2 (1972) Itô, K.: Extension of stochastic integrals, Proc. Intern. Symp. Stochastic on Differential Equations, K. Itô (ed.) (1978) 95 19, Kinokuniya. 7. Kuo, H.-H.: White Noise Distribution Theory, CRC Press, Kuo, H.-H.: Introduction to Stochastic Integration. Universitext (UTX), Springer, Kuo, H.-H.: The Itô calculus and white noise theory: A brief survey toward general stochastic integration, Communications on Stochastic Analysis 8, no. 1 (214) Kuo, H.-H. and Potthoff, J.: Anticipating stochastic integrals and stochastic differential equations; in: White Noise Analysis: Math. and Appl., T. Hida et al. (eds.), World Scientific (199) Kuo, H.-H., Sae-Tang, A., and Szozda, B.: A stochastic integral for adapted and instantly independent stochastic processes, in Advances in Statistics, Probability and Actuarial Science Vol. I, Stochastic Processes, Finance and Control: A Festschrift in Honour of Robert J. Elliott (eds.: Cohen, S., Madan, D., Siu, T. and Yang, H.), World Scientific, 212,

10 476 HUI-HSIUNG KUO AND KIMIAKI SAITÔ 12. León J. A. and Protter, P.: Some formulas for anticipative Girsanov transformations, in: Chaos Expansions, Multiple Wiener-Itô integrals and Their Applications, C. Houdré and V. Pérez-Abreu (eds.), CRC Press, Nualart, D. and Pardoux, E.: Stochastic calculus with anticipating integrands, Probab. Th. Rel. Fields 78 (1988) Pardoux, E. and Protter, P.: A two-sided stochastic integral and its calculus, Probab. Th. Rel. Fields 76 (1987) Russo, F. and Vallois, P.: Anticipative Stratonovich equation via Zvonkin method, Stochastic Processes and Related Topics (Siegmundsberg, 1994), , Stochastics Monogr., 1, Gordon and Breach, Yverdon, 1996, 16. Skorokhod, A. V.: On a generalization of a stochastic integral, Theory Probab. Appl. 2 (1975) Hui-Hsiung Kuo: Dept. of Mathematics, Louisiana State University, Baton Rouge, LA 783, USA. address: kuo@@math.lsu.edu Kimiaki Saitô: Department of Mathematics, Meijo University, Tenpaku, Nagoya , Japan address: ksaito@meijo-u.ac.jp

COVARIANCE IDENTITIES AND MIXING OF RANDOM TRANSFORMATIONS ON THE WIENER SPACE

COVARIANCE IDENTITIES AND MIXING OF RANDOM TRANSFORMATIONS ON THE WIENER SPACE Communications on Stochastic Analysis Vol. 4, No. 3 (21) 299-39 Serials Publications www.serialspublications.com COVARIANCE IDENTITIES AND MIXING OF RANDOM TRANSFORMATIONS ON THE WIENER SPACE NICOLAS PRIVAULT

More information

Stochastic Calculus (Lecture #3)

Stochastic Calculus (Lecture #3) Stochastic Calculus (Lecture #3) Siegfried Hörmann Université libre de Bruxelles (ULB) Spring 2014 Outline of the course 1. Stochastic processes in continuous time. 2. Brownian motion. 3. Itô integral:

More information

Man Kyu Im*, Un Cig Ji **, and Jae Hee Kim ***

Man Kyu Im*, Un Cig Ji **, and Jae Hee Kim *** JOURNAL OF THE CHUNGCHEONG MATHEMATICAL SOCIETY Volume 19, No. 4, December 26 GIRSANOV THEOREM FOR GAUSSIAN PROCESS WITH INDEPENDENT INCREMENTS Man Kyu Im*, Un Cig Ji **, and Jae Hee Kim *** Abstract.

More information

On Reflecting Brownian Motion with Drift

On Reflecting Brownian Motion with Drift Proc. Symp. Stoch. Syst. Osaka, 25), ISCIE Kyoto, 26, 1-5) On Reflecting Brownian Motion with Drift Goran Peskir This version: 12 June 26 First version: 1 September 25 Research Report No. 3, 25, Probability

More information

WHITE NOISE APPROACH TO FEYNMAN INTEGRALS. Takeyuki Hida

WHITE NOISE APPROACH TO FEYNMAN INTEGRALS. Takeyuki Hida J. Korean Math. Soc. 38 (21), No. 2, pp. 275 281 WHITE NOISE APPROACH TO FEYNMAN INTEGRALS Takeyuki Hida Abstract. The trajectory of a classical dynamics is detrmined by the least action principle. As

More information

The multidimensional Ito Integral and the multidimensional Ito Formula. Eric Mu ller June 1, 2015 Seminar on Stochastic Geometry and its applications

The multidimensional Ito Integral and the multidimensional Ito Formula. Eric Mu ller June 1, 2015 Seminar on Stochastic Geometry and its applications The multidimensional Ito Integral and the multidimensional Ito Formula Eric Mu ller June 1, 215 Seminar on Stochastic Geometry and its applications page 2 Seminar on Stochastic Geometry and its applications

More information

ON THE REGULARITY OF ONE PARAMETER TRANSFORMATION GROUPS IN BARRELED LOCALLY CONVEX TOPOLOGICAL VECTOR SPACES

ON THE REGULARITY OF ONE PARAMETER TRANSFORMATION GROUPS IN BARRELED LOCALLY CONVEX TOPOLOGICAL VECTOR SPACES Communications on Stochastic Analysis Vol. 9, No. 3 (2015) 413-418 Serials Publications www.serialspublications.com ON THE REGULARITY OF ONE PARAMETER TRANSFORMATION GROUPS IN BARRELED LOCALLY CONVEX TOPOLOGICAL

More information

Convergence at first and second order of some approximations of stochastic integrals

Convergence at first and second order of some approximations of stochastic integrals Convergence at first and second order of some approximations of stochastic integrals Bérard Bergery Blandine, Vallois Pierre IECN, Nancy-Université, CNRS, INRIA, Boulevard des Aiguillettes B.P. 239 F-5456

More information

ON A SDE DRIVEN BY A FRACTIONAL BROWNIAN MOTION AND WITH MONOTONE DRIFT

ON A SDE DRIVEN BY A FRACTIONAL BROWNIAN MOTION AND WITH MONOTONE DRIFT Elect. Comm. in Probab. 8 23122 134 ELECTRONIC COMMUNICATIONS in PROBABILITY ON A SDE DRIVEN BY A FRACTIONAL BROWNIAN MOTION AND WIT MONOTONE DRIFT BRAIM BOUFOUSSI 1 Cadi Ayyad University FSSM, Department

More information

Itô formula for generalized white noise functionals

Itô formula for generalized white noise functionals Itô formula for generalized white noise functionals Yuh-Jia Lee Department of Applied Mathematics National University of Kaohsiung Kaohsiung, TAIWAN 811 Worshop on IDAQP and their Applications 3-7 March,

More information

The Wiener Itô Chaos Expansion

The Wiener Itô Chaos Expansion 1 The Wiener Itô Chaos Expansion The celebrated Wiener Itô chaos expansion is fundamental in stochastic analysis. In particular, it plays a crucial role in the Malliavin calculus as it is presented in

More information

The concentration of a drug in blood. Exponential decay. Different realizations. Exponential decay with noise. dc(t) dt.

The concentration of a drug in blood. Exponential decay. Different realizations. Exponential decay with noise. dc(t) dt. The concentration of a drug in blood Exponential decay C12 concentration 2 4 6 8 1 C12 concentration 2 4 6 8 1 dc(t) dt = µc(t) C(t) = C()e µt 2 4 6 8 1 12 time in minutes 2 4 6 8 1 12 time in minutes

More information

A Barrier Version of the Russian Option

A Barrier Version of the Russian Option A Barrier Version of the Russian Option L. A. Shepp, A. N. Shiryaev, A. Sulem Rutgers University; shepp@stat.rutgers.edu Steklov Mathematical Institute; shiryaev@mi.ras.ru INRIA- Rocquencourt; agnes.sulem@inria.fr

More information

A DECOMPOSITION OF MULTIPLE WIENER INTEGRALS BY THE LÉVY PROCESS AND LÉVY LAPLACIAN

A DECOMPOSITION OF MULTIPLE WIENER INTEGRALS BY THE LÉVY PROCESS AND LÉVY LAPLACIAN Communications on Stochastic Analysis Vol. 2, No. 3 (2008) 459-467 Serials Publications www.serialspublications.com A DECOMPOSITION OF MULTIPLE WIENE INTEGALS BY THE LÉVY POCESS AND LÉVY LAPLACIAN ATSUSHI

More information

WHITE NOISE APPROACH TO THE ITÔ FORMULA FOR THE STOCHASTIC HEAT EQUATION

WHITE NOISE APPROACH TO THE ITÔ FORMULA FOR THE STOCHASTIC HEAT EQUATION Communications on Stochastic Analysis Vol. 1, No. 2 (27) 311-32 WHITE NOISE APPROACH TO THE ITÔ FORMULA FOR THE STOCHASTIC HEAT EQUATION ALBERTO LANCONELLI Abstract. We derive an Itô s-type formula for

More information

CHARACTERIZATION THEOREMS FOR DIFFERENTIAL OPERATORS ON WHITE NOISE SPACES

CHARACTERIZATION THEOREMS FOR DIFFERENTIAL OPERATORS ON WHITE NOISE SPACES Communications on Stochastic Analysis Vol. 7, No. 1 (013) 1-15 Serials Publications www.serialspublications.com CHARACTERIZATION THEOREMS FOR DIFFERENTIAL OPERATORS ON WHITE NOISE SPACES ABDESSATAR BARHOUMI

More information

Applications of Ito s Formula

Applications of Ito s Formula CHAPTER 4 Applications of Ito s Formula In this chapter, we discuss several basic theorems in stochastic analysis. Their proofs are good examples of applications of Itô s formula. 1. Lévy s martingale

More information

An essay on the general theory of stochastic processes

An essay on the general theory of stochastic processes Probability Surveys Vol. 3 (26) 345 412 ISSN: 1549-5787 DOI: 1.1214/1549578614 An essay on the general theory of stochastic processes Ashkan Nikeghbali ETHZ Departement Mathematik, Rämistrasse 11, HG G16

More information

Lecture 2. We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales.

Lecture 2. We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales. Lecture 2 1 Martingales We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales. 1.1 Doob s inequality We have the following maximal

More information

Stochastic Differential Equations.

Stochastic Differential Equations. Chapter 3 Stochastic Differential Equations. 3.1 Existence and Uniqueness. One of the ways of constructing a Diffusion process is to solve the stochastic differential equation dx(t) = σ(t, x(t)) dβ(t)

More information

A Class of Fractional Stochastic Differential Equations

A Class of Fractional Stochastic Differential Equations Vietnam Journal of Mathematics 36:38) 71 79 Vietnam Journal of MATHEMATICS VAST 8 A Class of Fractional Stochastic Differential Equations Nguyen Tien Dung Department of Mathematics, Vietnam National University,

More information

A REPRESENTATION FOR THE KANTOROVICH RUBINSTEIN DISTANCE DEFINED BY THE CAMERON MARTIN NORM OF A GAUSSIAN MEASURE ON A BANACH SPACE

A REPRESENTATION FOR THE KANTOROVICH RUBINSTEIN DISTANCE DEFINED BY THE CAMERON MARTIN NORM OF A GAUSSIAN MEASURE ON A BANACH SPACE Theory of Stochastic Processes Vol. 21 (37), no. 2, 2016, pp. 84 90 G. V. RIABOV A REPRESENTATION FOR THE KANTOROVICH RUBINSTEIN DISTANCE DEFINED BY THE CAMERON MARTIN NORM OF A GAUSSIAN MEASURE ON A BANACH

More information

Generalized Gaussian Bridges of Prediction-Invertible Processes

Generalized Gaussian Bridges of Prediction-Invertible Processes Generalized Gaussian Bridges of Prediction-Invertible Processes Tommi Sottinen 1 and Adil Yazigi University of Vaasa, Finland Modern Stochastics: Theory and Applications III September 1, 212, Kyiv, Ukraine

More information

Representing Gaussian Processes with Martingales

Representing Gaussian Processes with Martingales Representing Gaussian Processes with Martingales with Application to MLE of Langevin Equation Tommi Sottinen University of Vaasa Based on ongoing joint work with Lauri Viitasaari, University of Saarland

More information

Memoirs of My Research on Stochastic Analysis

Memoirs of My Research on Stochastic Analysis Memoirs of My Research on Stochastic Analysis Kiyosi Itô Professor Emeritus, Kyoto University, Kyoto, 606-8501 Japan It is with great honor that I learned of the 2005 Oslo Symposium on Stochastic Analysis

More information

On Integration-by-parts and the Itô Formula for Backwards Itô Integral

On Integration-by-parts and the Itô Formula for Backwards Itô Integral 11 On Integration-by-parts and the Itô Formula for... On Integration-by-parts and the Itô Formula for Backwards Itô Integral Jayrold P. Arcede a, Emmanuel A. Cabral b a Caraga State University, Ampayon,

More information

The Azéma-Yor Embedding in Non-Singular Diffusions

The Azéma-Yor Embedding in Non-Singular Diffusions Stochastic Process. Appl. Vol. 96, No. 2, 2001, 305-312 Research Report No. 406, 1999, Dept. Theoret. Statist. Aarhus The Azéma-Yor Embedding in Non-Singular Diffusions J. L. Pedersen and G. Peskir Let

More information

When is a Moving Average a Semimartingale?

When is a Moving Average a Semimartingale? 29 Barrett Lectures Ph.D.-student under supervision of Jan Pedersen, Thiele Centre, University of Aarhus, Denmark. 29 Barrett Lectures at The University of Tennessee: Stochastic Analysis and its Applications

More information

Topics in fractional Brownian motion

Topics in fractional Brownian motion Topics in fractional Brownian motion Esko Valkeila Spring School, Jena 25.3. 2011 We plan to discuss the following items during these lectures: Fractional Brownian motion and its properties. Topics in

More information

Exercises. T 2T. e ita φ(t)dt.

Exercises. T 2T. e ita φ(t)dt. Exercises. Set #. Construct an example of a sequence of probability measures P n on R which converge weakly to a probability measure P but so that the first moments m,n = xdp n do not converge to m = xdp.

More information

Independence of some multiple Poisson stochastic integrals with variable-sign kernels

Independence of some multiple Poisson stochastic integrals with variable-sign kernels Independence of some multiple Poisson stochastic integrals with variable-sign kernels Nicolas Privault Division of Mathematical Sciences School of Physical and Mathematical Sciences Nanyang Technological

More information

Self-intersection local time for Gaussian processes

Self-intersection local time for Gaussian processes Self-intersection local Olga Izyumtseva, olaizyumtseva@yahoo.com (in collaboration with Andrey Dorogovtsev, adoro@imath.kiev.ua) Department of theory of random processes Institute of mathematics Ukrainian

More information

Lecture 12. F o s, (1.1) F t := s>t

Lecture 12. F o s, (1.1) F t := s>t Lecture 12 1 Brownian motion: the Markov property Let C := C(0, ), R) be the space of continuous functions mapping from 0, ) to R, in which a Brownian motion (B t ) t 0 almost surely takes its value. Let

More information

(2m)-TH MEAN BEHAVIOR OF SOLUTIONS OF STOCHASTIC DIFFERENTIAL EQUATIONS UNDER PARAMETRIC PERTURBATIONS

(2m)-TH MEAN BEHAVIOR OF SOLUTIONS OF STOCHASTIC DIFFERENTIAL EQUATIONS UNDER PARAMETRIC PERTURBATIONS (2m)-TH MEAN BEHAVIOR OF SOLUTIONS OF STOCHASTIC DIFFERENTIAL EQUATIONS UNDER PARAMETRIC PERTURBATIONS Svetlana Janković and Miljana Jovanović Faculty of Science, Department of Mathematics, University

More information

Introduction to Diffusion Processes.

Introduction to Diffusion Processes. Introduction to Diffusion Processes. Arka P. Ghosh Department of Statistics Iowa State University Ames, IA 511-121 apghosh@iastate.edu (515) 294-7851. February 1, 21 Abstract In this section we describe

More information

BOOK REVIEW. Review by Denis Bell. University of North Florida

BOOK REVIEW. Review by Denis Bell. University of North Florida BOOK REVIEW By Paul Malliavin, Stochastic Analysis. Springer, New York, 1997, 370 pages, $125.00. Review by Denis Bell University of North Florida This book is an exposition of some important topics in

More information

Predicting the Time of the Ultimate Maximum for Brownian Motion with Drift

Predicting the Time of the Ultimate Maximum for Brownian Motion with Drift Proc. Math. Control Theory Finance Lisbon 27, Springer, 28, 95-112 Research Report No. 4, 27, Probab. Statist. Group Manchester 16 pp Predicting the Time of the Ultimate Maximum for Brownian Motion with

More information

Discrete approximation of stochastic integrals with respect to fractional Brownian motion of Hurst index H > 1 2

Discrete approximation of stochastic integrals with respect to fractional Brownian motion of Hurst index H > 1 2 Discrete approximation of stochastic integrals with respect to fractional Brownian motion of urst index > 1 2 Francesca Biagini 1), Massimo Campanino 2), Serena Fuschini 2) 11th March 28 1) 2) Department

More information

Eötvös Loránd University, Budapest. 13 May 2005

Eötvös Loránd University, Budapest. 13 May 2005 A NEW CLASS OF SCALE FREE RANDOM GRAPHS Zsolt Katona and Tamás F Móri Eötvös Loránd University, Budapest 13 May 005 Consider the following modification of the Barabási Albert random graph At every step

More information

Gaussian and Poisson White Noises with Related Characterization Theorems

Gaussian and Poisson White Noises with Related Characterization Theorems Contemporary Mathematics Gaussian and Poisson White Noises with Related Characterization Theorems Nobuhiro Asai, Izumi Kubo, and Hui-Hsiung Kuo Dedicated to Professor Leonard Gross on the occasion of his

More information

Bernardo D Auria Stochastic Processes /12. Notes. March 29 th, 2012

Bernardo D Auria Stochastic Processes /12. Notes. March 29 th, 2012 1 Stochastic Calculus Notes March 9 th, 1 In 19, Bachelier proposed for the Paris stock exchange a model for the fluctuations affecting the price X(t) of an asset that was given by the Brownian motion.

More information

Solution for Problem 7.1. We argue by contradiction. If the limit were not infinite, then since τ M (ω) is nondecreasing we would have

Solution for Problem 7.1. We argue by contradiction. If the limit were not infinite, then since τ M (ω) is nondecreasing we would have 362 Problem Hints and Solutions sup g n (ω, t) g(ω, t) sup g(ω, s) g(ω, t) µ n (ω). t T s,t: s t 1/n By the uniform continuity of t g(ω, t) on [, T], one has for each ω that µ n (ω) as n. Two applications

More information

Multiple points of the Brownian sheet in critical dimensions

Multiple points of the Brownian sheet in critical dimensions Multiple points of the Brownian sheet in critical dimensions Robert C. Dalang Ecole Polytechnique Fédérale de Lausanne Based on joint work with: Carl Mueller Multiple points of the Brownian sheet in critical

More information

RECENT PROGRESS ON THE WHITE NOISE APPROACH TO THE LÉVY LAPLACIAN

RECENT PROGRESS ON THE WHITE NOISE APPROACH TO THE LÉVY LAPLACIAN ECENT POGESS ON THE WHITE NOISE APPOACH TO THE LÉVY LAPLACIAN HUI-HSIUNG KUO Department of Mathematics Louisiana State University Baton ouge, LA 70808, USA E-mail: kuo@math.lsu.edu Let φ be a function

More information

STOCHASTIC CALCULUS JASON MILLER AND VITTORIA SILVESTRI

STOCHASTIC CALCULUS JASON MILLER AND VITTORIA SILVESTRI STOCHASTIC CALCULUS JASON MILLER AND VITTORIA SILVESTRI Contents Preface 1 1. Introduction 1 2. Preliminaries 4 3. Local martingales 1 4. The stochastic integral 16 5. Stochastic calculus 36 6. Applications

More information

On detection of unit roots generalizing the classic Dickey-Fuller approach

On detection of unit roots generalizing the classic Dickey-Fuller approach On detection of unit roots generalizing the classic Dickey-Fuller approach A. Steland Ruhr-Universität Bochum Fakultät für Mathematik Building NA 3/71 D-4478 Bochum, Germany February 18, 25 1 Abstract

More information

FOUNDATIONS OF MARTINGALE THEORY AND STOCHASTIC CALCULUS FROM A FINANCE PERSPECTIVE

FOUNDATIONS OF MARTINGALE THEORY AND STOCHASTIC CALCULUS FROM A FINANCE PERSPECTIVE FOUNDATIONS OF MARTINGALE THEORY AND STOCHASTIC CALCULUS FROM A FINANCE PERSPECTIVE JOSEF TEICHMANN 1. Introduction The language of mathematical Finance allows to express many results of martingale theory

More information

ELEMENTS OF STOCHASTIC CALCULUS VIA REGULARISATION. A la mémoire de Paul-André Meyer

ELEMENTS OF STOCHASTIC CALCULUS VIA REGULARISATION. A la mémoire de Paul-André Meyer ELEMENTS OF STOCHASTIC CALCULUS VIA REGULARISATION A la mémoire de Paul-André Meyer Francesco Russo (1 and Pierre Vallois (2 (1 Université Paris 13 Institut Galilée, Mathématiques 99 avenue J.B. Clément

More information

Maximum Process Problems in Optimal Control Theory

Maximum Process Problems in Optimal Control Theory J. Appl. Math. Stochastic Anal. Vol. 25, No., 25, (77-88) Research Report No. 423, 2, Dept. Theoret. Statist. Aarhus (2 pp) Maximum Process Problems in Optimal Control Theory GORAN PESKIR 3 Given a standard

More information

The Cameron-Martin-Girsanov (CMG) Theorem

The Cameron-Martin-Girsanov (CMG) Theorem The Cameron-Martin-Girsanov (CMG) Theorem There are many versions of the CMG Theorem. In some sense, there are many CMG Theorems. The first version appeared in ] in 944. Here we present a standard version,

More information

Rough paths methods 1: Introduction

Rough paths methods 1: Introduction Rough paths methods 1: Introduction Samy Tindel Purdue University University of Aarhus 216 Samy T. (Purdue) Rough Paths 1 Aarhus 216 1 / 16 Outline 1 Motivations for rough paths techniques 2 Summary of

More information

OPTIMAL SOLUTIONS TO STOCHASTIC DIFFERENTIAL INCLUSIONS

OPTIMAL SOLUTIONS TO STOCHASTIC DIFFERENTIAL INCLUSIONS APPLICATIONES MATHEMATICAE 29,4 (22), pp. 387 398 Mariusz Michta (Zielona Góra) OPTIMAL SOLUTIONS TO STOCHASTIC DIFFERENTIAL INCLUSIONS Abstract. A martingale problem approach is used first to analyze

More information

Some Tools From Stochastic Analysis

Some Tools From Stochastic Analysis W H I T E Some Tools From Stochastic Analysis J. Potthoff Lehrstuhl für Mathematik V Universität Mannheim email: potthoff@math.uni-mannheim.de url: http://ls5.math.uni-mannheim.de To close the file, click

More information

SOLUTIONS OF SEMILINEAR WAVE EQUATION VIA STOCHASTIC CASCADES

SOLUTIONS OF SEMILINEAR WAVE EQUATION VIA STOCHASTIC CASCADES Communications on Stochastic Analysis Vol. 4, No. 3 010) 45-431 Serials Publications www.serialspublications.com SOLUTIONS OF SEMILINEAR WAVE EQUATION VIA STOCHASTIC CASCADES YURI BAKHTIN* AND CARL MUELLER

More information

Quantum stochastic calculus applied to path spaces over Lie groups

Quantum stochastic calculus applied to path spaces over Lie groups Quantum stochastic calculus applied to path spaces over Lie groups Nicolas Privault Département de Mathématiques Université de La Rochelle Avenue Michel Crépeau 1742 La Rochelle, France nprivaul@univ-lr.fr

More information

HIDDEN MARKOV CHANGE POINT ESTIMATION

HIDDEN MARKOV CHANGE POINT ESTIMATION Communications on Stochastic Analysis Vol. 9, No. 3 (2015 367-374 Serials Publications www.serialspublications.com HIDDEN MARKOV CHANGE POINT ESTIMATION ROBERT J. ELLIOTT* AND SEBASTIAN ELLIOTT Abstract.

More information

Problem Sheet 1. You may assume that both F and F are σ-fields. (a) Show that F F is not a σ-field. (b) Let X : Ω R be defined by 1 if n = 1

Problem Sheet 1. You may assume that both F and F are σ-fields. (a) Show that F F is not a σ-field. (b) Let X : Ω R be defined by 1 if n = 1 Problem Sheet 1 1. Let Ω = {1, 2, 3}. Let F = {, {1}, {2, 3}, {1, 2, 3}}, F = {, {2}, {1, 3}, {1, 2, 3}}. You may assume that both F and F are σ-fields. (a) Show that F F is not a σ-field. (b) Let X :

More information

1. Stochastic Processes and filtrations

1. Stochastic Processes and filtrations 1. Stochastic Processes and 1. Stoch. pr., A stochastic process (X t ) t T is a collection of random variables on (Ω, F) with values in a measurable space (S, S), i.e., for all t, In our case X t : Ω S

More information

MA8109 Stochastic Processes in Systems Theory Autumn 2013

MA8109 Stochastic Processes in Systems Theory Autumn 2013 Norwegian University of Science and Technology Department of Mathematical Sciences MA819 Stochastic Processes in Systems Theory Autumn 213 1 MA819 Exam 23, problem 3b This is a linear equation of the form

More information

An Almost Sure Approximation for the Predictable Process in the Doob Meyer Decomposition Theorem

An Almost Sure Approximation for the Predictable Process in the Doob Meyer Decomposition Theorem An Almost Sure Approximation for the Predictable Process in the Doob Meyer Decomposition heorem Adam Jakubowski Nicolaus Copernicus University, Faculty of Mathematics and Computer Science, ul. Chopina

More information

Mutual Information for Stochastic Differential Equations*

Mutual Information for Stochastic Differential Equations* INFORMATION AND CONTROL 19, 265--271 (1971) Mutual Information for Stochastic Differential Equations* TYRONE E. DUNCAN Department of Computer, Information and Control Engineering, College of Engineering,

More information

Spatial Ergodicity of the Harris Flows

Spatial Ergodicity of the Harris Flows Communications on Stochastic Analysis Volume 11 Number 2 Article 6 6-217 Spatial Ergodicity of the Harris Flows E.V. Glinyanaya Institute of Mathematics NAS of Ukraine, glinkate@gmail.com Follow this and

More information

Supermodular ordering of Poisson arrays

Supermodular ordering of Poisson arrays Supermodular ordering of Poisson arrays Bünyamin Kızıldemir Nicolas Privault Division of Mathematical Sciences School of Physical and Mathematical Sciences Nanyang Technological University 637371 Singapore

More information

On Mean-Square and Asymptotic Stability for Numerical Approximations of Stochastic Ordinary Differential Equations

On Mean-Square and Asymptotic Stability for Numerical Approximations of Stochastic Ordinary Differential Equations On Mean-Square and Asymptotic Stability for Numerical Approximations of Stochastic Ordinary Differential Equations Rózsa Horváth Bokor and Taketomo Mitsui Abstract This note tries to connect the stochastic

More information

ESSENTIAL SETS FOR RANDOM OPERATORS CONSTRUCTED FROM AN ARRATIA FLOW

ESSENTIAL SETS FOR RANDOM OPERATORS CONSTRUCTED FROM AN ARRATIA FLOW Communications on Stochastic Analysis Vol., No. 3 (07) 30-3 Serials Publications www.serialspublications.com ESSENTIAL SETS FO ANDOM OPEATOS CONSTUCTED FOM AN AATIA FLOW A. A. DOOGOVTSEV AND IA. A. KOENOVSKA

More information

Contents. 1 Preliminaries 3. Martingales

Contents. 1 Preliminaries 3. Martingales Table of Preface PART I THE FUNDAMENTAL PRINCIPLES page xv 1 Preliminaries 3 2 Martingales 9 2.1 Martingales and examples 9 2.2 Stopping times 12 2.3 The maximum inequality 13 2.4 Doob s inequality 14

More information

Functional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals

Functional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals Functional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals Noèlia Viles Cuadros BCAM- Basque Center of Applied Mathematics with Prof. Enrico

More information

Random Fields: Skorohod integral and Malliavin derivative

Random Fields: Skorohod integral and Malliavin derivative Dept. of Math. University of Oslo Pure Mathematics No. 36 ISSN 0806 2439 November 2004 Random Fields: Skorohod integral and Malliavin derivative Giulia Di Nunno 1 Oslo, 15th November 2004. Abstract We

More information

n E(X t T n = lim X s Tn = X s

n E(X t T n = lim X s Tn = X s Stochastic Calculus Example sheet - Lent 15 Michael Tehranchi Problem 1. Let X be a local martingale. Prove that X is a uniformly integrable martingale if and only X is of class D. Solution 1. If If direction:

More information

1 Brownian Local Time

1 Brownian Local Time 1 Brownian Local Time We first begin by defining the space and variables for Brownian local time. Let W t be a standard 1-D Wiener process. We know that for the set, {t : W t = } P (µ{t : W t = } = ) =

More information

Regularity of the density for the stochastic heat equation

Regularity of the density for the stochastic heat equation Regularity of the density for the stochastic heat equation Carl Mueller 1 Department of Mathematics University of Rochester Rochester, NY 15627 USA email: cmlr@math.rochester.edu David Nualart 2 Department

More information

Malliavin calculus and central limit theorems

Malliavin calculus and central limit theorems Malliavin calculus and central limit theorems David Nualart Department of Mathematics Kansas University Seminar on Stochastic Processes 2017 University of Virginia March 8-11 2017 David Nualart (Kansas

More information

Estimates for the density of functionals of SDE s with irregular drift

Estimates for the density of functionals of SDE s with irregular drift Estimates for the density of functionals of SDE s with irregular drift Arturo KOHATSU-HIGA a, Azmi MAKHLOUF a, a Ritsumeikan University and Japan Science and Technology Agency, Japan Abstract We obtain

More information

STOCHASTIC ANALYSIS, CONTROLLED DYNAMICAL SYSTEMS AN APPLICATIONS

STOCHASTIC ANALYSIS, CONTROLLED DYNAMICAL SYSTEMS AN APPLICATIONS STOCHASTIC ANALYSIS, CONTROLLED DYNAMICAL SYSTEMS AN APPLICATIONS Jena, March 2015 Monique Jeanblanc, LaMME, Université d Évry-Val-D Essonne Enlargement of filtration in discrete time http://upload.wikimedia.org/wikipedia/commons/thumb/5/5a/dragon.jena.jpg/800px-dragon.jena.jpg

More information

The Uniform Integrability of Martingales. On a Question by Alexander Cherny

The Uniform Integrability of Martingales. On a Question by Alexander Cherny The Uniform Integrability of Martingales. On a Question by Alexander Cherny Johannes Ruf Department of Mathematics University College London May 1, 2015 Abstract Let X be a progressively measurable, almost

More information

Branching Processes II: Convergence of critical branching to Feller s CSB

Branching Processes II: Convergence of critical branching to Feller s CSB Chapter 4 Branching Processes II: Convergence of critical branching to Feller s CSB Figure 4.1: Feller 4.1 Birth and Death Processes 4.1.1 Linear birth and death processes Branching processes can be studied

More information

Numerical methods for solving stochastic differential equations

Numerical methods for solving stochastic differential equations Mathematical Communications 4(1999), 251-256 251 Numerical methods for solving stochastic differential equations Rózsa Horváth Bokor Abstract. This paper provides an introduction to stochastic calculus

More information

Strong Solutions and a Bismut-Elworthy-Li Formula for Mean-F. Formula for Mean-Field SDE s with Irregular Drift

Strong Solutions and a Bismut-Elworthy-Li Formula for Mean-F. Formula for Mean-Field SDE s with Irregular Drift Strong Solutions and a Bismut-Elworthy-Li Formula for Mean-Field SDE s with Irregular Drift Thilo Meyer-Brandis University of Munich joint with M. Bauer, University of Munich Conference for the 1th Anniversary

More information

LECTURE 2: LOCAL TIME FOR BROWNIAN MOTION

LECTURE 2: LOCAL TIME FOR BROWNIAN MOTION LECTURE 2: LOCAL TIME FOR BROWNIAN MOTION We will define local time for one-dimensional Brownian motion, and deduce some of its properties. We will then use the generalized Ray-Knight theorem proved in

More information

Krzysztof Burdzy University of Washington. = X(Y (t)), t 0}

Krzysztof Burdzy University of Washington. = X(Y (t)), t 0} VARIATION OF ITERATED BROWNIAN MOTION Krzysztof Burdzy University of Washington 1. Introduction and main results. Suppose that X 1, X 2 and Y are independent standard Brownian motions starting from 0 and

More information

arxiv: v1 [math.pr] 20 May 2018

arxiv: v1 [math.pr] 20 May 2018 arxiv:180507700v1 mathr 20 May 2018 A DOOB-TYE MAXIMAL INEQUALITY AND ITS ALICATIONS TO VARIOUS STOCHASTIC ROCESSES Abstract We show how an improvement on Doob s inequality leads to new inequalities for

More information

VITA of Yaozhong HU. A. List of submitted papers

VITA of Yaozhong HU. A. List of submitted papers VITA of Yaozhong HU A. List of submitted papers 1. (with G. Rang) Parameter Estimation For Stochastic Hamiltonian Systems Driven By Fractional Brownian Motions. 2. (with G. Rang) Identification of the

More information

A NOTE ON STOCHASTIC INTEGRALS AS L 2 -CURVES

A NOTE ON STOCHASTIC INTEGRALS AS L 2 -CURVES A NOTE ON STOCHASTIC INTEGRALS AS L 2 -CURVES STEFAN TAPPE Abstract. In a work of van Gaans (25a) stochastic integrals are regarded as L 2 -curves. In Filipović and Tappe (28) we have shown the connection

More information

Brownian Motion and Stochastic Calculus

Brownian Motion and Stochastic Calculus ETHZ, Spring 17 D-MATH Prof Dr Martin Larsson Coordinator A Sepúlveda Brownian Motion and Stochastic Calculus Exercise sheet 6 Please hand in your solutions during exercise class or in your assistant s

More information

Solving the Poisson Disorder Problem

Solving the Poisson Disorder Problem Advances in Finance and Stochastics: Essays in Honour of Dieter Sondermann, Springer-Verlag, 22, (295-32) Research Report No. 49, 2, Dept. Theoret. Statist. Aarhus Solving the Poisson Disorder Problem

More information

ON THE FIRST TIME THAT AN ITO PROCESS HITS A BARRIER

ON THE FIRST TIME THAT AN ITO PROCESS HITS A BARRIER ON THE FIRST TIME THAT AN ITO PROCESS HITS A BARRIER GERARDO HERNANDEZ-DEL-VALLE arxiv:1209.2411v1 [math.pr] 10 Sep 2012 Abstract. This work deals with first hitting time densities of Ito processes whose

More information

Information and Credit Risk

Information and Credit Risk Information and Credit Risk M. L. Bedini Université de Bretagne Occidentale, Brest - Friedrich Schiller Universität, Jena Jena, March 2011 M. L. Bedini (Université de Bretagne Occidentale, Brest Information

More information

STATISTICS 385: STOCHASTIC CALCULUS HOMEWORK ASSIGNMENT 4 DUE NOVEMBER 23, = (2n 1)(2n 3) 3 1.

STATISTICS 385: STOCHASTIC CALCULUS HOMEWORK ASSIGNMENT 4 DUE NOVEMBER 23, = (2n 1)(2n 3) 3 1. STATISTICS 385: STOCHASTIC CALCULUS HOMEWORK ASSIGNMENT 4 DUE NOVEMBER 23, 26 Problem Normal Moments (A) Use the Itô formula and Brownian scaling to check that the even moments of the normal distribution

More information

TWO RESULTS ON MULTIPLE STRATONOVICH INTEGRALS

TWO RESULTS ON MULTIPLE STRATONOVICH INTEGRALS Statistica Sinica 71997, 97-9 TWO RESULTS ON MULTIPLE STRATONOVICH INTEGRALS A. Budhiraja and G. Kallianpur University of North Carolina, Chapel Hill Abstract: Formulae connecting the multiple Stratonovich

More information

Law of total probability and Bayes theorem in Riesz spaces

Law of total probability and Bayes theorem in Riesz spaces Law of total probability and Bayes theorem in Riesz spaces Liang Hong Abstract. This note generalizes the notion of conditional probability to Riesz spaces using the order-theoretic approach. With the

More information

POWERS OF AN INFINITE DIMENSIONAL BROWNIAN MOTION ASSOCIATED WITH THE PRODUCT OF DISTRIBUTIONS

POWERS OF AN INFINITE DIMENSIONAL BROWNIAN MOTION ASSOCIATED WITH THE PRODUCT OF DISTRIBUTIONS Communications on Stochastic Analysis Vol. 8, No. 3 (04 343-364 Serials Publications www.serialspublications.com POWERS OF AN INFINITE DIMENSIONAL BROWNIAN MOTION ASSOCIATED WITH THE PRODUCT OF DISTRIBUTIONS

More information

Riemann-Stieltjes integrals and fractional Brownian motion

Riemann-Stieltjes integrals and fractional Brownian motion Riemann-Stieltjes integrals and fractional Brownian motion Esko Valkeila Aalto University, School of Science and Engineering, Department of Mathematics and Systems Analysis Workshop on Ambit processes,

More information

Weak solutions of mean-field stochastic differential equations

Weak solutions of mean-field stochastic differential equations Weak solutions of mean-field stochastic differential equations Juan Li School of Mathematics and Statistics, Shandong University (Weihai), Weihai 26429, China. Email: juanli@sdu.edu.cn Based on joint works

More information

On Stochastic Adaptive Control & its Applications. Bozenna Pasik-Duncan University of Kansas, USA

On Stochastic Adaptive Control & its Applications. Bozenna Pasik-Duncan University of Kansas, USA On Stochastic Adaptive Control & its Applications Bozenna Pasik-Duncan University of Kansas, USA ASEAS Workshop, AFOSR, 23-24 March, 2009 1. Motivation: Work in the 1970's 2. Adaptive Control of Continuous

More information

Stochastic Shear Thickening Fluids: Strong Convergence of the Galerkin Approximation and the Energy Equality 1

Stochastic Shear Thickening Fluids: Strong Convergence of the Galerkin Approximation and the Energy Equality 1 Stochastic Shear Thickening Fluids: Strong Convergence of the Galerkin Approximation and the Energy Equality Nobuo Yoshida Contents The stochastic power law fluids. Terminology from hydrodynamics....................................

More information

16. D. C. Struppa and C. Turrini, Hyperfunctions and boundary values ofholomorphic functions, Nieuw Arch. Wisk. 4 (1986),

16. D. C. Struppa and C. Turrini, Hyperfunctions and boundary values ofholomorphic functions, Nieuw Arch. Wisk. 4 (1986), 170 BOOK REVIEWS 16. D. C. Struppa and C. Turrini, Hyperfunctions and boundary values ofholomorphic functions, Nieuw Arch. Wisk. 4 (1986), 91-118. JOHN HORVÂTH UNIVERSITY OF MARYLAND, COLLEGE PARK BULLETIN

More information

Comment on Weak Convergence to a Matrix Stochastic Integral with Stable Processes

Comment on Weak Convergence to a Matrix Stochastic Integral with Stable Processes Comment on Weak Convergence to a Matrix Stochastic Integral with Stable Processes Vygantas Paulauskas Department of Mathematics and Informatics, Vilnius University Naugarduko 24, 03225 Vilnius, Lithuania

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 9 10/2/2013. Conditional expectations, filtration and martingales

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 9 10/2/2013. Conditional expectations, filtration and martingales MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 9 10/2/2013 Conditional expectations, filtration and martingales Content. 1. Conditional expectations 2. Martingales, sub-martingales

More information

Some SDEs with distributional drift Part I : General calculus. Flandoli, Franco; Russo, Francesco; Wolf, Jochen

Some SDEs with distributional drift Part I : General calculus. Flandoli, Franco; Russo, Francesco; Wolf, Jochen Title Author(s) Some SDEs with distributional drift Part I : General calculus Flandoli, Franco; Russo, Francesco; Wolf, Jochen Citation Osaka Journal of Mathematics. 4() P.493-P.54 Issue Date 3-6 Text

More information

Part III Stochastic Calculus and Applications

Part III Stochastic Calculus and Applications Part III Stochastic Calculus and Applications Based on lectures by R. Bauerschmidt Notes taken by Dexter Chua Lent 218 These notes are not endorsed by the lecturers, and I have modified them often significantly

More information