JOURNAL OF THE CHUNGCHEONG MATHEMATICAL SOCIETY Volume 19, No. 4, December 26 GIRSANOV THEOREM FOR GAUSSIAN PROCESS WITH INDEPENDENT INCREMENTS Man Kyu Im*, Un Cig Ji **, and Jae Hee Kim *** Abstract. A characterization of Gaussian process with independent increments in terms of the support of covariance operator is established. We investigate the Girsanov formula for a Gaussian process with independent increments. 1. Introduction Since the stochastic calculus for standard Brownian motion initiated by K. Itô, the stochastic calculi for several stochastic processes have been extensively developed with wide applications to physics, mathematical finance, engineering, biology etc, in [2, 4, 5, 9, 1, 11, 12 and the references cited therein. Recently, the stochastic calculus for Gaussian processes have been developed in [1, 3, 6, 7. The main purpose of this paper is two folds. We first establish a characterization of Gaussian process with independent increments in terms of the support of covariance operator. Secondly, we study the Girsanov theorem for a Gaussian process with independent increments. We study a Gaussian process with its characterization and representation. Then we study stochastic calculus for the Gaussian process, and we investigate the Girsanov formula for the Gaussian process. As applications of the Girsanov theorem, the studies of financial model and nonlinear filtering problems for the Gaussian process are now in progress, see [3, 6, 8, 9, 1. This paper is organized as follows: In Section 2 we study a characterization of a Gaussian process with independent increments. In Section Received November 14, 26. 2 Mathematics Subject Classification: Primary 6G15, 6G51, 6G44. Key words and phrases: Gaussian process, stochastic integral, Itô formula, Girsanov theorem. **This work was supported by Chungbuk National University Grant in 26.
384 Man Kyu Im, Un Cig Ji, and Jae Hee Kim 3 we recall the stochastic integral and Itô formula for the Gaussian process. In Section 4 we establish the Girsanov theorem for the Gaussian process. 2. Gaussian process with independent increments Let H = L 2 (R +, dt) and K a strictly positive selfadjoint operator with domain D H containing all indicator functions, where R + = [, ). Let (Ω, F, P ) be a complete probability space and let B = {B K,t } t be a Gaussian process with mean E[B K,t = α(t) and covariance function (2.1) E[(B K,t α(t))(b K,s α(s)) = K1 [,t, 1 [,s, s, t, where α is of bounded variation on closed intervals with α() =. Theorem 2.1. For any s t, supp(k1 [s,t ) [s, t if and only if the Gaussian process {B K,t } t has independent increments. (2.2) Proof. If supp(kf) supp(f) for any f D, then by (2.1) we have E[(B K,t B K,s )(B K,v B K,u ) = E[(B K,t B K,s )E[(B K,v B K,u ), s t u v. Hence {B K,t } t has independent increments. Conversely, if {B K,t } t has independent increments, then (2.2) holds and so K1 [s,t, 1 [u,v = for s t u v. The rest of the proof is straightforward. Theorem 2.2. Let T > and K a positive selfadjoint bounded operator on L 2 ([, T ) such that supp(kf) supp(f) for any f L 2 ([, T ). Then there exists a strictly increasing absolutely continuous function β on [, T for any T > such that the multiplication operator M β = K, where β (t) = dβ(t)/dt for t < T. Proof. Let β(t) = K1 [,t, 1 [,t for t T. Then β(t) is a strictly increasing function, in fact, for any = t < t 1 < < t n T we have β(t i ) β(t i 1 ) = K1 [,ti, 1 [,ti K1 [,ti 1, 1 [,ti 1 = ti t i 1 K1 [ti 1,t i (u)du. Therefore, β is absolutely continuous on [, T. Since M 1A K = KM 1A for any measurable set A [, T, KM f = M f K for any bounded measurable function f. On the other hand, the algebra of all multiplication operators of bounded measurable functions is maximal abelian
Girsanov Theorem for Gaussian Process 385 and so K = M g for some g L ([, T ). Therefore we have β(t) = M g 1 [,t, 1 [,t = t g(s)ds which implies that g = β. From now on, we assume that supp(k1 [s,t ) [s, t for any s t and put β(t) = K1 [,t, 1 [,t for t, and {B K,t } t has a continuous version. For each t we put A K,t = B K,t α(t). Then {A K,t } t is a Gaussian process such that A K,t has the normal distribution with mean and variance β(t). 3. Stochastic integral and Itô formula Let {B K,t } t be a Gaussian process. For each t, let B t be the σ algebra generated by {B K,s : s t} and then we write B K,t = (B K,t, B t ) for t. Definition 3.1. Let T > and M G = M G [, T the class of functions f : [, T Ω R such that (i) the map (t, ω) f(t, ω) is B([, T ) F measurable, (ii) for each t T, f(t, ) is B t measurable, (iii) E f(t, ω) 2 d (β(t) + α (t)) <. A function φ M G is called an elementary function if it has the form: (3.1) φ(t, ω) = j e j (ω)1 [tj,t j+1 )(t), where e j is B tj -measurable. For the elementary function φ given as in (3.1), we define (3.2) T φ(t, ω)db K,t (ω) = j e j (ω) ( B K,tj+1 B K,tj ). The integral defined as in (3.2) is called the stochastic integral of φ with respect to the Gaussian process {B K,t } t. We first note that if f M G, we can choose elementary functions φ n M G such that (3.3) E f φ n 2 d(β + α )(t) as n,
386 Man Kyu Im, Un Cig Ji, and Jae Hee Kim for the proof, we refer to [6. Therefore, for each f M G, the stochastic integral with respect to the process {B K,t } t is defined by T f(t, ω)db K,t (ω) = lim n T φ n (t, ω)db K,t (ω), where {φ n } is a sequence of elementary functions in M G given as in (3.3) and the limit exists in L 2 (Ω). Then we can prove the following elementary inequality: for any f M G we have [ ( T ) 2 E f(t, ω)db K,t (ω) 2E f(t, ω) 2 dβ(t) + 2V T (α)e f(t, ω) 2 d α (t), where V T (α) is the total variation of α over [, T, see [6. A diffusion process for a Gaussian process {B K,t } t is a stochastic process X t given by dx t = u(t, ω)dt + v(t, ω)db K,t, t, [ T where v M G [, T and u is B t -adapted with E u(s, ω)2 ds <. Theorem 3.2 ([6). Let X t be a diffusion process given by dx t = udt + vdb K,t. Let g C 2 (R + R). Then Y t = g(t, X t ) is again a diffusion process and satisfies the following stochastic differential equation: dy t = g t (t, X t)dt + g x (t, X t)dx t + 1 2 g(t, X t ) 2 x 2 (dx t ) 2, where (dx t ) 2 = (dx t )(dx t ) is computed according to the rules: dt dt = dt db K,t = db K,t dt =, db K,t db K,t = β(t). 4. The Girsanov theorem In this section we study the Girsanov Theorem for a Gaussian process with independent increments. We start with a theorem for the existence and uniqueness of (strong) solution of a stochastic differential equation. For the proof, we refer to [6.
Girsanov Theorem for Gaussian Process 387 Theorem 4.1. Let T > and b(, ) : [, T R R, σ(, ) : [, T R R be measurable functions satisfying b(t, x) + σ(t, x) C(t)(1 + x ), x R, t [, T ; b(t, x) b(t, y) + σ(t, x) σ(t, y) C(t) x y, x, y R, t [, T for some bounded measurable function C. Let Z be a random variable which is independent of the σ-algebra B generated by {B K,s : s } and such that E[ Z 2 <. Then the stochastic differential equation dx t = b(t, X t )dt + σ(t, X t )db K,t, X = Z, t T has a unique t-continuous solution X t with the property that X t is adapted to the filtration Bt Z generated by Z and B t and E X t 2 d(ρ + β + α )(t) <, where ρ is the identity function on R +. Theorem 4.2. Let {X t } t be a continuous stochastic process on the probability space (Ω, F, P ). Then the following are equivalent: (i) {X t } t is a Gaussian process with independent increments, and mean and variance β(t). (ii) {X t } t is a martingale with respect to the filtration {B t } t of which the quadratic variation on [, t is β(t). Proof. The proof of (i) (ii) is straightforward. To prove the converse we assume that (ii) is satisfied. The proof is a simple modification of the proof of Theorem 6.1 in [5. For each t, we put Z t = e iuxt, u R. Then by applying Itô formula (see Theorem 5.1 in [5) we obtain that t (4.1) Z t = Z s + iuz v dx v u2 Z v dβ(v), < s t. s 2 s On the other hand, it is obvious that [ t E iuz v dx v B s = P a.s. From (4.1) for each A B s we have E[e iu(x t X s ) 1 A = P (A) u2 2 which implies that s t t s E[e iu(x v X s ) 1 A dβ(v) E[e iu(xt Xs) 1 A = P (A) exp{ u2 2 (β(t) β(s))}, A B s.
388 Man Kyu Im, Un Cig Ji, and Jae Hee Kim Hence we prove that for s < t E[exp{iu(X t X s )} B s = exp{ u2 (β(t) β(s))} P a.s. 2 We conclude that X t X s is independent of B s and that normally distributed with mean and variance β(t) β(s). Lemma 4.3 ([1). Let µ and ν be two probability measures on a measurable space (Ω, F) such that dν(ω) = Z(ω)dµ(ω) for some Z L 1 (Ω, µ). Let X be a random variable on (Ω, F) such that E ν [ X = X(ω) Z(ω)dµ(ω) <. Let H be a σ-algebra with H F. Then Ω E ν [X H E µ [Z H = E µ [ZX H P a.s. From now on, we assume that the mean function α of the Gaussian process {B K,t } t is absolutely continuous. Theorem 4.4. Let {Y t } t T be a diffusion process of the form: dy t = v(t)dt + db K,t, t T, Y = for a B t -adapted process {v(t)} t T, where T is a given constant and {B K,t } t is a Gaussian process with mean function α(t) and the variance function β(t). Assume that [ ( ) 1 T (v(s) + α (s)) 2 (4.2) E exp 2 β ds <, (s) where E = E P is the expectation with respect to P. Put ( t v(s) + α (s) (4.3) M t = exp β da K,s 1 t (s) 2 β (s) (v(s) + α (s)) 2 ) ds for t T, where da K,t = db K,t α (t)dt. Then {Y t } t T is a Gaussian process on (Ω, B T, Q) with mean and variance β(t), where the probability measure Q is defined by (4.4) dq(ω) = M T (ω)dp (ω). In fact, Q(Ω) = E P [M T = 1 since {M t } t T is a martingale.
Girsanov Theorem for Gaussian Process 389 Proof. For the proof we use similar arguments used in the proof of Theorem 8.6.3 in [1. For each t T we put K t = M t Y t. Then by using Itô formula we have dk t = M t dy t + Y t dm t + dy t dm [ ( t v(t) + α ) (t) = M t 1 Y t β da K,t. (t) Therefore, K t is a martingale with respect to P and so, by Lemma 4.3, for s < t T we have E Q [Y t B s = E[M ty t B s E[M t B s = E[K t B s M s = K s M s = Y s which implies that {Y t } t T is a martingale with respect to the probability Q. Therefore, E Q [Y t = E Q [Y = for any t T. On the other hand, since the process {Y t } t has independent increments, by applying Itô formula we have E [ [ Yt 2 t = 2E Y s dy s + β(t) = β(t), t T. Therefore, {Y t } t T is a martingale with respect to the filtration {B t } and the probability measure Q of which the quadratic variation on [, t is β(t). Hence by Theorem 4.2, {Y t } t T is a Gaussian process on (Ω, B T, Q) with mean and variance β(t). Remark 4.5. Note that (4.2) is a sufficient condition to be that {M t } is a martingale and so, in general, Theorem 4.4 is true with the assumption that {M t } is a martingale. Theorem 4.6. Let {Y t } t T be a diffusion process of the form: (4.5) dy t = u(t)dt + θ(t)db K,t, t T for a B t -adapted process {u(t)} t T. Suppose there exist processes {v(t)} t T M G and {η(t)} t T M G such that (4.6) θ(t)v(t) = u(t) η(t), t T. Assume that (4.2) holds. Let Q be the measure defined as in (4.4). Then (4.7) BK,t = t v(s)ds + B K,t, t T is a Gaussian process on (Ω, B T, Q) with mean and variance β(t), and the process {Y t } t T has the following stochastic integral representation: dy t = η(t)dt + θ(t)d B K,t, t T.
39 Man Kyu Im, Un Cig Ji, and Jae Hee Kim Proof. It is obvious from Theorem 4.4 that { B K,t } t T is a Gaussian process on (Ω, B T, Q) with mean and variance β(t). On the other hand, by (4.5), (4.7) and (4.6) we have dy t = u(t)dt + θ(t)(d B K,t v(t)dt) = η(t)dt + θ(t)d B K,t. Hence we complete the proof. Theorem 4.7. Let {X t } t T and {Y t } t T be diffusion processes of the forms: dx t = b(x t )dt + σ(x t )db K,t, dy t = [γ(t) + b(y t )dt + σ(y t )db K,t, t T, X = Y = x R, where the functions b and σ satisfy the conditions of Theorem 4.1 and γ(t) M G. Suppose there exists a stochastic process {v(t)} t T such that σ(y t ) ( v(t) + α (t) ) = γ(t), t T. Assume that (4.2) holds. Then the process {Y t } t T has the following stochastic integral representation: ( (4.8) dy t = b(y t )dt + σ(y t ) d B ) K,t + dα(t), t T, where { B K,t } t T is given as in (4.7). Proof. By applying Theorem 4.6 to the case θ(t) = σ(y t ), u(t) = γ(t) + b(y t ) and η(t) = b(y t ) + σ(y t )α (t), we have the representation (4.8). Remark 4.8. Let {X t } t T and {Y t } t T be the diffusion processes given as in Theorem 4.7, and Q a probability measure defined as in (4.4). For each t T, put B K,t = α(t) + B K,t for the Gaussian process { B K,t } given as in (4.7). Then {B K,t } is a Gaussian process on (Ω, B T, Q) with mean α(t) and variance β(t), and (4.8) becomes dy t = b(y t )dt + σ(y t )db K,t, t T. Therefore, by the uniqueness of a solution of a stochastic differential equation in Theorem 4.1, the Q-law of {Y t } t T is the same as the P -law of {X t } t T.
Girsanov Theorem for Gaussian Process 391 References [1 E. Alòs, O. Mazet, and D. Nualart, Stochastic calculus with respect to Gaussian processes, Ann. Probab. 29 (21), 766 81. [2 E. Alòs and D. Nualart, Stochastic integration with respect to the fractional Brownian motion, Stoch. Stoch. Rep. 75 (23), 129 152. [3 T. S. Chung and U. C. Ji, Gaussian white noise and applications to finance, in Quantum Information and Complexity (T. Hida, K. Saitô and Si Si, Eds.), World Scientific, 24, pp. 179 21. [4 T. Duncan, Y. Hu, and B. Pasik-Duncan, Stochastic calculus for fractional Brownian motion I, Theory, SIAM J. Control Optim. 38 (2), 582 612. [5 N. Ikeda and S. Watanabe, Stochastic Differential Equations and Diffusion Process, (North-Holland/Kodansha, 1989). [6 M. K. Im, U. C. Ji, and J. H. Kim, Stochastic calculus for generalized Wiener process and linear filtering, Preprint, 25. [7 U. C. Ji and N. Obata, Unitarity of generalized Fourier-Gauss transforms, Stoch. Anal. Appl. 24 (26), 733 751. [8 G. Kallianpur and R. L. Karandikar, White Noise Theory of Prediction, Filtering and Smoothing, (Gordon & Breach Science Publishers, New York, 1988). [9 R. S. Liptser and A. N. Shiryayev, Statistics of Random Processes, (2nd ed.) (Springer Verlag, 21). [1 B. Øksendal, Stochastic Differential Equations: An Introduction with Applications, (Springer-Verlag, Berlin, 1998). [11 F. Russo and P. Vallois, Stochastic calculus with respect to continuous finite quadratic variation processes, Stoch. Stoch. Rep. 7 (2), 1 4. [12 J. Yeh, Stochastic Processes and the Wiener Integral, (Marcel Deckker, New York, 1973). * Department of Mathematics Hannam University Daejon 36-791, Republic of Korea E-mail: mki@mail.hannam.ac.kr ** Department of Mathematics Research Institute of Mathematical Finance Chungbuk National University Cheongju 361-763, Republic of Korea E-mail: uncigji@cbucc.chungbuk.ac.kr *** Department of Mathematics Hanyang University Seoul 133-791, Republic of Korea