A BAYES FORMULA FOR NON-LINEAR FILTERING WITH GAUSSIAN AND COX NOISE
|
|
- Jasper Stafford
- 6 years ago
- Views:
Transcription
1 A BAYES FOMULA FO NON-LINEA FILTEING WITH GAUSSIAN AND COX NOISE V. MANDEKA, T. MEYE-BANDIS, AND F. POSKE Abstract. A Bayes type formula is derived for the non-linear filter where the observation contains both general Gaussian noise as well as Cox noise whose jump intensity depends on the signal. This formula extends the well know Kallianpur-Striebel formula in the classical non-linear filter setting. We also discuss Zakai type equations for both the unnormalized conditional distribution as well as unnormalized conditional density in case the signal is a Markovian jump diffusion.. Introduction The general filtering setting can be described as follows. Assume a partially observable process (X, Y = (X t, Y t t T 2 defined on a probability space (Ω, F, P. The real valued process X t stands for the unobservable component, referred to as the signal process or system process, whereas Y t is the observable part, called observation process. Thus information about X t can only be obtained by extracting the information about X that is contained in the observation Y t in a best possible way. In filter theory this is done by determining the conditional distribution of X t given the information σ-field Ft Y generated by Y s, s t. Or stated in an equivalent way, the objective is to compute the optimal filter as the conditional expectation for a rich enough class of functions f. E P f(x t F Y t In the classical non-linear filter setting, the dynamics of the observation process Y t is supposed to follow the following Itô process dy t = h(t, X t dt + dw t, where W t is a Brownian motion independent of X. Under certain conditions on the drift h(t, X t (see KS, K, Kallianpur and Striebel derived a Bayes type formula for the conditional distribution expressed in terms of the so called unnormalized conditional distribution. In the special case when the dynamics of the signal follows an Itô diffusion dy t = b(t, X t dt + σ(t, X t db t, for a second Brownian motion B t, Zakai (Z showed under certain conditions that the unormalized conditional density is the solution of an associated stochastic partial differential equation, the so called Zakai equation. Date: April 28, 29. Key words and phrases. Non-linear Filtering, Gaussian noise, Cox noise, Bayes formula, Kallianpur- Striebel formula, Zakai euqation.
2 2 MANDEKA, MEYE-BANDIS, AND POSKE In this paper we extend the classical filter model to the following more general setting. For a general signal process X we suppose the observation model is given as t (. Y t = β(t, X + G t + ς N λ (dt, dς, where G t is a general Gaussian process with zero mean and continuous covariance function (s, t, s, t T, that is independent of the signal process X. Let Ft Y (respectively Ft X denote the σ-algebra generated by {Y s, s t} (respectively {X s, s t} augmented by the null-sets. Define the filtration (F t t T through F t := FT X F t Y. Then we assume that the process t L t := ς N λ (dt, dς is a pure jump F t -semimartingale determined through the integer valued random measure N λ that has an F t -predictable compensator of the form µ(dt, dς, ω = λ(t, X, ςdtν(dς for a Lévy measure ν and a functional λ(t, X(ω, ς. In particular, G t and L t are independent. The function β :, T,T is such that β(t, is Ft X -measurable and β(, X(ω is in H( for almost all ω, where H( denotes the Hilbert space generated by (s, t (see Section 2. The observation dynamics consists thus of an information drift of the signal disturbed by some Gaussian noise plus a pure jump part whose jump intensity depends on the signal. Note that a jump process of the form given above is also referred to as Cox process. The objective of the paper is in a first step to extend the Kallianpur-Striebel Bayes type formula to the generalized filter setting from above. When there are no jumps present in the observation dynamics (. the corresponding formula has been developped in (MM. We will extend their way of reasoning to the situation including Cox noise. In a second step we then derive a Zakai type measure valued stochastic differential equations for the unnormalized conditional distribution of the filter. For this purpose we assume the signal process X to be a Markov process with generator O t := L t + B t given as L t f(x := b(t, x x f(x + 2 σ2 (t, x xx f(x B t f(x := {f(x + γ(t, xς f(x x f(xγ(t, xς} υ(dς, with the coefficients b(t, x, σ(t, x, andγ(t, x and f(x being in C 2 ( for every t. Here, C 2 ( is the space of continuous functions with compact support and bounded derivatives up to order 2. Further, we develop a Zakai type stochastic parabolic integro partial differential equation for the unnormalized conditional density, given it exists. In the case the dynamics of X does not contain any jumps and the Gaussian noise G t in the observation
3 is Brownian motion, the corresponding Zakai equation was also studied in (MP. For further information on Zakai equations in a semimartingale setting we also refer to (G and (G2. The remaining part of the paper is organized as follows. in Section 2 we briefly recall some theory of reproducing kernel Hilbert spaces. In Section 3 we obtain the Kallianpur- Striebel formula, before we discuss the Zakai type equations in Section eproducing Kernel Hilbert Space and Stochastic Processes A Hilbert space H consisting of real valued functions on some set T is said to be a reproducing kernel Hilbert space (KHS, if there exists a function K on T T with the following two properties: for every t in T and g in H, (i K(, t H, (ii (g(, K(, t = g(t. (The reproducing property K is called the reproducing kernel of H. The following basic properties can be found in A. ( If a reproducing kernel exists, then it is unique. (2 If K is the reproducing kernel of a Hilbert space H, then {K(, t, t T} spans H. (3 If K is the reproducing kernel of a Hilbert space H, then it is nonnegative definite in the sense that for all t,..., t n in T and a,..., a n n K(t i, t j a i a j. i,j= The converse of (3, stated in Theorem 2. below, is fundamental towards understanding the KHS representation of Gaussian processes. A proof of the theorem can be found in A. Theorem 2. (E. H. Moore. A symmetric nonnegative definite function K on T T generates a unique Hilbert space, which we denote by H(K or sometimes by H(K, T, of which K is the reproducing kernel. Now suppose K(s, t, s, t T, is a nonnegative definite function. Then, by Theorem 2., there is a KHS, H(K, T, with K as its reproducing kernel. If we restrict K to T T where T T, then K is still a nonnegative definite function. Hence K restricted to T T will also correspond to a reproducing kernel Hilbert space H(K, T of functions defined on T. The following result from (A; pp. 35 explains the relationship between these two. Theorem 2.2. Suppose K T, defined on T T, is the reproducing kernel of the Hilbert space H(K T with the norm. Let T T, and K T be the restriction of K T on T T. Then H(K T consists of all f in H(K T restricted to T. Further, for such a restriction f H(K T the norm f H(KT is the minimum of f H(KT for all f H(K T whose restriction to T is f. If K(s, t is the covariance function for some zero mean process Z t, t T, then, by Theorem 2., there exists a unique KHS, H(K, T, for which K is the reproducing kernel. It is also easy to see (e.g., see Theorem 3D, P that there exists a congruence (linear, 3
4 4 MANDEKA, MEYE-BANDIS, AND POSKE one-to-one, inner product preserving map between H(K and sp L2 {Z t, t T} which takes K(, t to Z t. Let us denote by Z, h sp L2 {Z t, t T}, the image of h H(K, T under the congruence. We conclude the section with an important special case. 2.. A useful example. Suppose the stochastic process Z t is a Gaussian process given by Z t = t F (t, udw u, t T, where t F 2 (t, udu < for all t T and W u is Brownian motion. Then the covariance function (2. K(s, t E(Z s Z t = and the corresponding KHS is given by { (2.2 H(K = g : g(t = t t s F (t, uf (s, udu, } F (t, ug (udu, t T for some (necessarily unique g sp L2 {F (t,,t (, t T }, with the inner product where g (s = s (g, g 2 H(K = T g (ug 2(udu, F (s, ug (udu and g 2 (s = s F (s, ug 2(udu. For t T, by taking K(, t to be F (t,,t (, we see, from (2. and (2.2, that K(, t H(K. To check the reproducing property suppose h(t = t F (t, uh (u du H(K. Then (h, K(, t H(K = T h (uk(, t du = t h (uf (t, u du = h(t. Also, in this case, it is very easy to check (cf. P2, Theorem 4D that the congruence between H(K and sp L2 {Z t, t T} is given by (2.3 Z, g = T g (udw u. 3. The Filter Setting and a Bayes Formula Assume a partially observable process (X, Y = (X t, Y t t T 2 defined on a probability space (Ω, F, P. The real valued process X t stands for the unobservable component, referred to as the signal process, whereas Y t is the observable part, called observation process. In particular, we assume that the dynamics of the observation process is given as follows: t (3. Y t = β(t, X + G t + ς N λ (dt, dς, where
5 G t is a Gaussian process with zero mean and continuous covariance function (s, t, s, t T, that is independent of the signal process X. The function β :, T,T is such that β(t, is Ft X -measurable and β(, X(ω is in H( for almost all ω, where H( denotes the Hilbert space generated by (s, t (see Section 2. Let Ft Y (respectively Ft X denote the σ-algebra generated by {Y s, s t} (respectively {X s, s t} augmented by the null-sets. Define the filtration (F t t T through F t := FT X F t Y. Then we assume that the process t L t := ς N λ (dt, dς is a pure jump F t -semimartingale determined through the integer valued random measure N λ that has an F t -predictable compensator of the form µ(dt, dς, ω = λ(t, X, ςdtν(dς for a Lévy measure ν and a functional λ(t, X(ω, ς. The functional λ(t, X, ς is assumed to be strictly positive and such that T (3.2 log 2 (λ(s, X, ς µ(ds, dς < a.s. T log 2 (λ(s, X, ς ds ν(dς < a.s. (3.3 and { t ( Λ t := exp log λ(s, X, ς t ( ( + log λ(s, X, ς Ñ λ (ds, dς } λ(s, X, ς + µ(ds, dς is a well defined F t -martingale. Here Ñλ(ds, dς stands for the compensated jump measure Ñ λ (ds, dς := N λ (ds, dς µ(dt, dς. emark 3.. Note that the specific from of the predictable compensator µ(dt, dς, ω implies that L t is a process with conditionally independent increments with respect to the σ-algebra F X T, i.eėpf(lt LsA F X T = E P f(l t L s F X T E P A F X T, for all bounded measurable functions f, A F s, and s < t T (see for example Th. 6.6 in JS. Also, it follows that the processes G is independent from the random measure N λ (ds, dς. Given a Borel measurable function f, our non-linear filtering problem then comes down to determine the least square estimate of f(x t, given the observations up to time t. In other words, the problem consists in evaluating the optimal filter (3.4 E P f(x t F t Y. 5
6 6 MANDEKA, MEYE-BANDIS, AND POSKE In this section we want to derive a Bayes formula for the optimal filter (3.4 by an extension of the reference measure method presented in MM for the purely Gaussian case. For this purpose, define for each t T with β( = β(, X { Λ t := exp G, β t } 2 β 2 t Then the main tool is the following extension of Theorem 3. in MM Lemma 3.2. Define dq := Λ t Λ t dp. Then Q t is a probability measure, and under Q t we have that Y t = G t + L t, where G s = β(s, X + G s, s t, is a Gaussian process with zero mean and covariance function, L s, s t, is a pure jump Lévy process with Lévy measure ν, and the process X s, s T has the same distribution as under P. Further, the processes G, L and X are independent under Q t. Proof. Fix t T. First note that since β( H( almost surely, we have by Theorem 2.2 that β,t H(; t almost surely. Further, by the independence of the Gaussian process G from X and from the random measure N λ (ds, dς it follows that E P Λ t Λ t = E P E P Λ t F X T E P Λ t F X T. Since for f H(; t the random variable G, f t is Gaussian with zero mean and variance f 2 t, it follows again by the independence of G from X and the martingale property of Λ t that E P Λ t Λ t =, and Q t is a probability measure. Now take s,..., s m t, r,..., r p t, t,..., t n T and real numbers λ,..., λ m, γ,..., γ p, α,..., α n and consider the joint characteristic function E Qt e i P n j= α jx tj +i P m i= λ ig e si +i P p k= γ k(l rk L rk = E P e i P n j= α jx tj +i P m i= λ ig e si +i P p k= γ k(l rk L rk Λ t Λ t = E P e i P n j= α jx tj E P e i P m i= λ ig e si Λ t FT X E P e i P p k= γ k(l rk L rk Λ t FT X. Here, for computational convenience, the part of the characteristic function that concerns L is formulated in terms of increments of L (where we set r =. Now, as in Theorem 3. in MM, we get by the independence of G from X that E P e i P m i= λ i e G si Λ t F X T = e P m i,l= λ iλ l (s i,s l, which is the characteristic function of a Gaussian process with mean zero and covariance function. Further, by the conditional independent increments of L we get like in the proof of Th. 6.6 in JS that e u E r δ(s,x,ς N e λ (ds,dς u P F X T = e r (e δ(s,x,ς δ(s,x,ς µ(dt,dς
7 for r u T. So that for one increment one obtains E P e iγ(lu Lr Λ t FT X { u ( ( = E P exp iγς + log Ñ λ (ds, dς r λ(s, X, ς u ( ( } + iγς + log r λ(s, X, ς λ(s, X, ς + µ(dt, dς FT X { u ( = E P exp (e iγς+log λ(s,x,ς iγς log µ(dt, dς r λ(s, X, ς u ( ( } + iγς + log r λ(s, X, ς λ(s, X, ς + µ(dt, dς FT X { u } = E P exp (e iγς+log λ(s,x,ς λ(t, X, ςdtν(dς FT X r λ(s, X, ς { ( = exp (u r e iγς } ν(dς. The generalization to the sum of increments is straightforward and one obtains the characteristic function of the finite dimensional distribution of a Lévy process (of finite variation: { E P e i P p p k= γ k(l rk L rk Λ t FT X ( = exp (r k r k e iγ k ς } ν(dς. All together we end up with k= E Qt e i P n j= α jx tj +i P m i= λ i e G si +i P p k= γ k(l rk L rk = E P e i P n j= α jx tj e P m i,l= λ iλ l (s i,s l ep p k= (r k r k (e iγ k ς ν(dς, which completes the proof. emark 3.3. Note that in case G is Brownian motion Lemma 3.2 is just the usual Girsanov theorem for Brownian motion and random measures. In this case, it follows from Cameron-Martin s result and the fact that X is independent of G that Λ t Λ t is a martingale and dq is a probability measure. Now, the inverse adon-nikodym derivative dp dq t = (Λ t (Λ t is Q t -a.s. by condition (3.2 and an argument like in (MM, p. 857 given through { t (Λ t = exp log (λ(s, X, ς Ñ(ds, dς t } + (log (λ(s, X, ς λ(s, X, ς + ds ν(dς { (Λ t = exp G, β t } 2 β 2 t. 7
8 8 MANDEKA, MEYE-BANDIS, AND POSKE Here Ñ(ds, dς := N λ (ds, dς dtν(dς is now a compesnsated Poisson random measure under Q t. Then we have by the Bayes formula for conditional expectation for any FT X -measurable integrable function g(t, X E P g(t, X F Y E Qt g(t, X(Λt (Λ t = t Ft Y E Qt (Λt (Λ t Ft Y. ( From Lemma 3.2 we know that the processes Gs, (L s s t, and (X s s T are s t independent under Q t and that the distribution of X is the same under Q t as under P. Hence conditional expectations of the form E Qt φ(x, G, L Ft Y can be computed as E Qt φ(x, G, L Ft Y (ω = φ(x(ˆω, G(ω, L(ω Q t (dˆω Ω = φ(x(ˆω, G(ω, L(ω P(dˆω = φ(x(ˆω, G(ω, L(ω. Ω where (ω, ˆω Ω Ω and the index ˆP denotes integration with respect to ˆω. Consequently, we get the following Bayes formula for the optimal filter Theorem 3.4. Under the above specified conditions, for any FT X -measurable integrable function g(t, X E P g(t, X F Y Ω t = g(t, X(ˆωα t(ω, ˆωα t(ω, ˆω P(dˆω Ω α t(ω, ˆωα t (ω, ˆω P(dˆω = g(t, X(ˆωα t (ω, ˆωα t(ω, ˆω α t (ω, ˆωα t (ω, ˆω, where { t α t (ω, ˆω = exp log (λ(s, X(ˆω, ς Ñ(ω, ds, dς t } + (log (λ(s, X(ˆω, ς λ(s, X(ˆω, ς + ds ν(dς { α t(ω, ˆω = exp G(ω, β(, ˆω t } 2 β(, ˆω 2 t. 4. Zakai type equations Using the Bayes formula from above we now want to proceed further in deriving a Zakai type equations for the unnormalized filter. This equation is basic in order to obtain the filter recursively. To this end we have to impose certain restrictions on both the signal process and the Gaussian part of the observation process. egarding the signal process X, we assume its dynamics to be Markov. More precisely, we consider the parabolic integro-differential operator O t := L t + B t, where L t f(x := b(t, x x f(x + 2 σ2 (t, x xx f(x
9 B t f(x := {f(x + γ(t, xς f(x x f(xγ(t, xς} υ(dς, for f C 2(. Here, C2 ( is the space of continuous functions with compact support and bounded derivatives up to order 2. Further, we suppose that b(t,, σ(t,, and γ(t, are in C 2 ( for every t and that υ(dς is a Lévy measure with second moment. The signal process X t, t T, is then assumed to be a solution of the martingale problem corresponding to O t, i.e. f(x t t (O u f(x u du is a Ft X -martingale with respect to P for every f C 2(. Further, we restrict the Gaussian process G of the observation process in (3. to belong to the special case presented in Section 2., i.e. G t = t F (t, s dw s, where W t is Brownian motion and F (t, s is a deterministic function such that t F 2 (t, s ds, t T. Note that this type of processes both includes Ornstein-Uhlenbeck processes as well as fractional Brownian motion. Then β(t, X will be of the form Further, with β(t, X = W t := t t F (t, sh(s, X s ds. h(s, X s ds + W t we get G, β t = t h(s, X s d W s and β 2 t = t h2 (s, X s ds, and α t(ω, ˆω in Theorem 3.4 becomes { t α t(ω, ˆω = exp h(s, X s (ˆω d W s (ω t } h 2 (s, X s (ˆω ds. 2 Note that in this case W s, s t, is a Brownian motion under Q t. For f C 2( we now define the unnormalized filter V t(f = V t (f(ω by V t (f(ω := f(x t (ˆωα t (ω, ˆωα t(ω, ˆω P(dˆω = f(xt (ˆωα t (ω, ˆωα t(ω, ˆω. Ω Then this unnormalized filter obeys the following dynamics Theorem 4.. (Zakai equation I Under the above specified assumptions, the unnormalized filter V t (f satisfies the equation ( ( (4. dv t (f( (ω = V t O t f( (ω dt + V t h(t, f( (ω d W t (ω ( (4.2 + V t (λ(t,, ς f( (ω Ñ(ω, dt, dς. 9
10 MANDEKA, MEYE-BANDIS, AND POSKE Proof. Set g t (ˆω := f(x T (ˆω T t (O s f(x s (ˆω ds. Then, by our assumptions on the coefficients b, σ, γ and on the Lévy measure υ(dς, we have g t < C for some constant C. Since f(x t t O tf(x s ds is a martingale we obtain (4.3 g t F X(ˆω t = f(x t, t T. If we denote Γ t (ω, ˆω := α t (ω, ˆωα t(ω, ˆω, then, because Γ t (ω, ˆω is F X(ˆω t -measurable for each ω, equation (4.3 implies that By definition of g t, V t (f = f(x t (ˆωΓ t (ω, ˆω = = g t (ˆωΓ t (ω, ˆω g t (ˆωΓ t (ω, ˆω F X(ˆω t dg t (ˆω = (O t f(x t (ˆω dt. Also, Γ t = Γ t (ω, ˆω is the Doléans-Dade solution of the following linear SDE dγ t = h(t, X t (ˆωΓ t d W t (ω + (λ(t, X t (ˆω, ς Γ t Ñ(ω, dt, dς. So we get t g t (ˆωΓ t = g (ˆωΓ + (O s f(x s (ˆωΓ s ds t h(s, X s (ˆωg s (ˆωΓ s d W s (ω + + t (λ(s, X s (ˆω, ς g s (ˆωΓ s Ñ(ω, ds, dς. The first term on the right hand side equals f(x, and for the second one we can invoke Fubini s theorem to get t t t (O s f(x s (ˆωΓ s ds = (O s f(x s (ˆωΓ s ds = V s (O s f( (ω ds. For the third term we employ the stochastic Fubini theorem for Brownian motion (see for example 5.4 in LS in order to get t h(s, X s (ˆωg s (ˆωΓ s d W s (ω = = t t h(s, X s (ˆωg s (ˆωΓ s d W s (ω h(s, X s (ˆωΓ s g s (ˆω Fs X(ˆω d W s (ω
11 = = t t h(s, X s (ˆωΓ s f(x s (ˆω d W s (ω V s (h(s, f( (ω d W s (ω. Further, one easily sees that the analogue stochastic Fubini theorem for compensated Poisson random measures holds, and we get analogously for the last term t (λ(s, X s (ˆω, ς g s (ˆωΓ s Ñ(ω, ds, dς t ( = (λ(s,, ς f( (ω Ñ(ω, ds, dς, which completes the proof. V s If one further assume that the filter has a so called unormalized conditional density u(t, x then we can derive a stochastic integro-pde determining u(t, x which for the Brownian motion case was first established in Z and usually is referred to as Zakai equation. Definition 4.2. We say that a process u(t, x = u(ω, t, x is the unnormalized conditional density of the filter if (4.4 V t (f(ω = f(xu(ω, t, x dx for all bounded continuous functions f :. From now on we restrict the intergo part B t of the operator O t to be the one of a pure jump Lévy process, i.e. γ =, and we assume the initial value X (ω of the signal process to posses a density denoted by ξ(x. Then the following holds Theorem 4.3. (Zakai equation II Suppose the unnormalized conditional density u(t, x of our filter exists. Then, provided a solution exists, u(t, x solves the following stochastic integro-pde (4.5 du(t, x = Ot u(t, xdt + h(t, xu(t, xd W t (ω + (λ(t, x, ς u(t, xñ(ω, dt, dς u(, x = ξ(x. Here Ot := L t + Bt is the adjoint operator of O t given through L t f(x := x (b(t, xf(x + 2 ( xx σ 2 (t, xf(x Bt f(x := {f(x ς f(x + x f(xς} υ(dς, for f C 2 (. For sufficient conditions on the coefficients under which there exists a classical solution of (4.5 see for example MP; in M the existence of solutions in a generalized sense of stochastic distributions is treated.
12 2 MANDEKA, MEYE-BANDIS, AND POSKE Proof. By (4. and (4.4 we have for all f C ( t f(xu(t, x dx = f(xξ(x dx + u(s, xosf(x dxds + t t Now, using integration by parts, we get (4.6 u(s, xl sf(x dx = + u(s, xh(s, xf(x dxd W s (ω u(s, x (λ(s, x, ς f(x dxñ(ω, ds, dς f(xo su(s, x dx. Further it holds again integration by parts and by substitution that (4.7 u(s, xbsf(x dx = f(xbsu(s, x dx. Fubini together with (4.6 and (4.7 then yields ( t f(xu(t, x dx = f(xξ(x dx + f(x + Since this holds for all f C + ( t f(x Osu(s, x ds dx u(s, xh(s, x d W s (ω dx ( t f(x u(s, x (λ(s, x, ς Ñ(ω, ds, dς dx ( we get (4.5. Acknowledgements: This work was started when V. Mandrekar was at the CMA,University of Oslo. He thanks the CMA for support and Prof. Øksendal for interesting discussions and hospitality. eferences A N. Aronszajn (95: Theory of reproducing kernels, Trans. Amer. Math. Soc., 68, pp G Grigelionis B.: Stochastic non-linear filtering equations and semimartingales. In Lecture Notes Math. 972, eds Mitter, S.K., Moro, A., Springer-Verlag, Berlin (982. G2 Grigelionis B.: Stochastic evolution equations and densities of the conditional distributions,lecture notes in Control and Info. Sci., 983, p. 49. JS Jacod, J., Shiryaev, A.N. (22: Limit Theorems for Stochastic Processes. Springer, 2. ed., Berlin Heidelberg New York. K G. Kallianpur (98: Stochastic Filtering Theory. Springer Verlag, New York, Heidelberg, Berlin, 98. KS G. Kallianpur, C. Striebel (968: Estimations of stochastic processes : Arbitrary system process with additive white noise observation errors. Ann. Math. Statis., 39 (968, pp LS Liptser.S., Shiryaev A.N. (977: Statistics of andom Processes. Vol., Springer-Verlag, NY, 977. MM P. Mandal and V. Mandrekar (2: A Bayes formula for Gaussian noise processes and its applications, SIAM J. of Control and Optimization
13 3 M Meyer-Brandis T. (25: Stochastic Feynman-Kac Equations associated to Lévy-Itô Diffusions, Stochastic Analysis and Applications, to appear. MP Meyer-Brandis, T., Proske, F.: Explicit Solution of a Non-Linear Filtering Problem for Lévy Processes with Application to Finance, Appl. Math. Optim. 5 (24, pp P E. Parzen(96: egression Analysis of Continuous Parameter Time Series, in Proc. 4th Berkeley Sympos. Math. Statis. and Prob., Vol. I, Univ. California Press, Berkeley, Calif., 96, pp P2 E. Parzen(967: Statistical inference on time series by Hilbert space methods, I, in Time Series Analysis Papers, Holden-Day, San Francisco, Cambridge, London, Amsterdam, 967, pp Z Zakai, M. (969: On the optimal filtering of diffusion processes. Z. Wahrsch. verw. Geb., No., , (969. (Vidyadhar Mandrekar, Department of Statistics and Probability, Michigan State University, East Lansing, MI48824, USA (Thilo Meyer-Brandis, Center of Mathematics for Applications, University of Oslo, P.O. Box 53, Blindern, N 36 Oslo, Norway, (Frank Proske, Center of Mathematics for Applications, University of Oslo, P.O. Box 53, Blindern, N 36 Oslo, Norway,
Man Kyu Im*, Un Cig Ji **, and Jae Hee Kim ***
JOURNAL OF THE CHUNGCHEONG MATHEMATICAL SOCIETY Volume 19, No. 4, December 26 GIRSANOV THEOREM FOR GAUSSIAN PROCESS WITH INDEPENDENT INCREMENTS Man Kyu Im*, Un Cig Ji **, and Jae Hee Kim *** Abstract.
More informationPROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS
PROBABILITY: LIMIT THEOREMS II, SPRING 218. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please
More informationTools of stochastic calculus
slides for the course Interest rate theory, University of Ljubljana, 212-13/I, part III József Gáll University of Debrecen Nov. 212 Jan. 213, Ljubljana Itô integral, summary of main facts Notations, basic
More informationStochastic Differential Equations
CHAPTER 1 Stochastic Differential Equations Consider a stochastic process X t satisfying dx t = bt, X t,w t dt + σt, X t,w t dw t. 1.1 Question. 1 Can we obtain the existence and uniqueness theorem for
More informationA Short Introduction to Diffusion Processes and Ito Calculus
A Short Introduction to Diffusion Processes and Ito Calculus Cédric Archambeau University College, London Center for Computational Statistics and Machine Learning c.archambeau@cs.ucl.ac.uk January 24,
More informationFiltrations, Markov Processes and Martingales. Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition
Filtrations, Markov Processes and Martingales Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition David pplebaum Probability and Statistics Department,
More informationOn Reflecting Brownian Motion with Drift
Proc. Symp. Stoch. Syst. Osaka, 25), ISCIE Kyoto, 26, 1-5) On Reflecting Brownian Motion with Drift Goran Peskir This version: 12 June 26 First version: 1 September 25 Research Report No. 3, 25, Probability
More informationPROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS
PROBABILITY: LIMIT THEOREMS II, SPRING 15. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please
More informationTopics in fractional Brownian motion
Topics in fractional Brownian motion Esko Valkeila Spring School, Jena 25.3. 2011 We plan to discuss the following items during these lectures: Fractional Brownian motion and its properties. Topics in
More informationOn the martingales obtained by an extension due to Saisho, Tanemura and Yor of Pitman s theorem
On the martingales obtained by an extension due to Saisho, Tanemura and Yor of Pitman s theorem Koichiro TAKAOKA Dept of Applied Physics, Tokyo Institute of Technology Abstract M Yor constructed a family
More informationRandom Fields: Skorohod integral and Malliavin derivative
Dept. of Math. University of Oslo Pure Mathematics No. 36 ISSN 0806 2439 November 2004 Random Fields: Skorohod integral and Malliavin derivative Giulia Di Nunno 1 Oslo, 15th November 2004. Abstract We
More informationThe Cameron-Martin-Girsanov (CMG) Theorem
The Cameron-Martin-Girsanov (CMG) Theorem There are many versions of the CMG Theorem. In some sense, there are many CMG Theorems. The first version appeared in ] in 944. Here we present a standard version,
More informationApplications of Ito s Formula
CHAPTER 4 Applications of Ito s Formula In this chapter, we discuss several basic theorems in stochastic analysis. Their proofs are good examples of applications of Itô s formula. 1. Lévy s martingale
More informationOn a class of stochastic differential equations in a financial network model
1 On a class of stochastic differential equations in a financial network model Tomoyuki Ichiba Department of Statistics & Applied Probability, Center for Financial Mathematics and Actuarial Research, University
More informationHJB equations. Seminar in Stochastic Modelling in Economics and Finance January 10, 2011
Department of Probability and Mathematical Statistics Faculty of Mathematics and Physics, Charles University in Prague petrasek@karlin.mff.cuni.cz Seminar in Stochastic Modelling in Economics and Finance
More informationS. Lototsky and B.L. Rozovskii Center for Applied Mathematical Sciences University of Southern California, Los Angeles, CA
RECURSIVE MULTIPLE WIENER INTEGRAL EXPANSION FOR NONLINEAR FILTERING OF DIFFUSION PROCESSES Published in: J. A. Goldstein, N. E. Gretsky, and J. J. Uhl (editors), Stochastic Processes and Functional Analysis,
More information{σ x >t}p x. (σ x >t)=e at.
3.11. EXERCISES 121 3.11 Exercises Exercise 3.1 Consider the Ornstein Uhlenbeck process in example 3.1.7(B). Show that the defined process is a Markov process which converges in distribution to an N(0,σ
More informationMathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( )
Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio (2014-2015) Etienne Tanré - Olivier Faugeras INRIA - Team Tosca November 26th, 2014 E. Tanré (INRIA - Team Tosca) Mathematical
More informationBrownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539
Brownian motion Samy Tindel Purdue University Probability Theory 2 - MA 539 Mostly taken from Brownian Motion and Stochastic Calculus by I. Karatzas and S. Shreve Samy T. Brownian motion Probability Theory
More informationHarnack Inequalities and Applications for Stochastic Equations
p. 1/32 Harnack Inequalities and Applications for Stochastic Equations PhD Thesis Defense Shun-Xiang Ouyang Under the Supervision of Prof. Michael Röckner & Prof. Feng-Yu Wang March 6, 29 p. 2/32 Outline
More informationRepresentations of Gaussian measures that are equivalent to Wiener measure
Representations of Gaussian measures that are equivalent to Wiener measure Patrick Cheridito Departement für Mathematik, ETHZ, 89 Zürich, Switzerland. E-mail: dito@math.ethz.ch Summary. We summarize results
More informationA Concise Course on Stochastic Partial Differential Equations
A Concise Course on Stochastic Partial Differential Equations Michael Röckner Reference: C. Prevot, M. Röckner: Springer LN in Math. 1905, Berlin (2007) And see the references therein for the original
More information1 Brownian Local Time
1 Brownian Local Time We first begin by defining the space and variables for Brownian local time. Let W t be a standard 1-D Wiener process. We know that for the set, {t : W t = } P (µ{t : W t = } = ) =
More informationBernardo D Auria Stochastic Processes /10. Notes. Abril 13 th, 2010
1 Stochastic Calculus Notes Abril 13 th, 1 As we have seen in previous lessons, the stochastic integral with respect to the Brownian motion shows a behavior different from the classical Riemann-Stieltjes
More informationNonlinear Filtering Revisited: a Spectral Approach, II
Nonlinear Filtering Revisited: a Spectral Approach, II Sergey V. Lototsky Center for Applied Mathematical Sciences University of Southern California Los Angeles, CA 90089-3 sergey@cams-00.usc.edu Remijigus
More informationGeneralized Gaussian Bridges of Prediction-Invertible Processes
Generalized Gaussian Bridges of Prediction-Invertible Processes Tommi Sottinen 1 and Adil Yazigi University of Vaasa, Finland Modern Stochastics: Theory and Applications III September 1, 212, Kyiv, Ukraine
More informationLecture 4: Introduction to stochastic processes and stochastic calculus
Lecture 4: Introduction to stochastic processes and stochastic calculus Cédric Archambeau Centre for Computational Statistics and Machine Learning Department of Computer Science University College London
More informationLecture 12. F o s, (1.1) F t := s>t
Lecture 12 1 Brownian motion: the Markov property Let C := C(0, ), R) be the space of continuous functions mapping from 0, ) to R, in which a Brownian motion (B t ) t 0 almost surely takes its value. Let
More informationA NOTE ON STOCHASTIC INTEGRALS AS L 2 -CURVES
A NOTE ON STOCHASTIC INTEGRALS AS L 2 -CURVES STEFAN TAPPE Abstract. In a work of van Gaans (25a) stochastic integrals are regarded as L 2 -curves. In Filipović and Tappe (28) we have shown the connection
More informationVerona Course April Lecture 1. Review of probability
Verona Course April 215. Lecture 1. Review of probability Viorel Barbu Al.I. Cuza University of Iaşi and the Romanian Academy A probability space is a triple (Ω, F, P) where Ω is an abstract set, F is
More informationfor all f satisfying E[ f(x) ] <.
. Let (Ω, F, P ) be a probability space and D be a sub-σ-algebra of F. An (H, H)-valued random variable X is independent of D if and only if P ({X Γ} D) = P {X Γ}P (D) for all Γ H and D D. Prove that if
More informationStochastic Differential Equations.
Chapter 3 Stochastic Differential Equations. 3.1 Existence and Uniqueness. One of the ways of constructing a Diffusion process is to solve the stochastic differential equation dx(t) = σ(t, x(t)) dβ(t)
More informationA Change of Variable Formula with Local Time-Space for Bounded Variation Lévy Processes with Application to Solving the American Put Option Problem 1
Chapter 3 A Change of Variable Formula with Local Time-Space for Bounded Variation Lévy Processes with Application to Solving the American Put Option Problem 1 Abstract We establish a change of variable
More informationOn the quantiles of the Brownian motion and their hitting times.
On the quantiles of the Brownian motion and their hitting times. Angelos Dassios London School of Economics May 23 Abstract The distribution of the α-quantile of a Brownian motion on an interval [, t]
More informationBernardo D Auria Stochastic Processes /12. Notes. March 29 th, 2012
1 Stochastic Calculus Notes March 9 th, 1 In 19, Bachelier proposed for the Paris stock exchange a model for the fluctuations affecting the price X(t) of an asset that was given by the Brownian motion.
More informationEXISTENCE OF NEUTRAL STOCHASTIC FUNCTIONAL DIFFERENTIAL EQUATIONS WITH ABSTRACT VOLTERRA OPERATORS
International Journal of Differential Equations and Applications Volume 7 No. 1 23, 11-17 EXISTENCE OF NEUTRAL STOCHASTIC FUNCTIONAL DIFFERENTIAL EQUATIONS WITH ABSTRACT VOLTERRA OPERATORS Zephyrinus C.
More informationThe Wiener Itô Chaos Expansion
1 The Wiener Itô Chaos Expansion The celebrated Wiener Itô chaos expansion is fundamental in stochastic analysis. In particular, it plays a crucial role in the Malliavin calculus as it is presented in
More informationEstimates for the density of functionals of SDE s with irregular drift
Estimates for the density of functionals of SDE s with irregular drift Arturo KOHATSU-HIGA a, Azmi MAKHLOUF a, a Ritsumeikan University and Japan Science and Technology Agency, Japan Abstract We obtain
More informationLecture 21 Representations of Martingales
Lecture 21: Representations of Martingales 1 of 11 Course: Theory of Probability II Term: Spring 215 Instructor: Gordan Zitkovic Lecture 21 Representations of Martingales Right-continuous inverses Let
More informationWhite noise generalization of the Clark-Ocone formula under change of measure
White noise generalization of the Clark-Ocone formula under change of measure Yeliz Yolcu Okur Supervisor: Prof. Bernt Øksendal Co-Advisor: Ass. Prof. Giulia Di Nunno Centre of Mathematics for Applications
More informationOPTIMAL SOLUTIONS TO STOCHASTIC DIFFERENTIAL INCLUSIONS
APPLICATIONES MATHEMATICAE 29,4 (22), pp. 387 398 Mariusz Michta (Zielona Góra) OPTIMAL SOLUTIONS TO STOCHASTIC DIFFERENTIAL INCLUSIONS Abstract. A martingale problem approach is used first to analyze
More informationA Note on the Central Limit Theorem for a Class of Linear Systems 1
A Note on the Central Limit Theorem for a Class of Linear Systems 1 Contents Yukio Nagahata Department of Mathematics, Graduate School of Engineering Science Osaka University, Toyonaka 560-8531, Japan.
More informationI forgot to mention last time: in the Ito formula for two standard processes, putting
I forgot to mention last time: in the Ito formula for two standard processes, putting dx t = a t dt + b t db t dy t = α t dt + β t db t, and taking f(x, y = xy, one has f x = y, f y = x, and f xx = f yy
More informationMultivariate Generalized Ornstein-Uhlenbeck Processes
Multivariate Generalized Ornstein-Uhlenbeck Processes Anita Behme TU München Alexander Lindner TU Braunschweig 7th International Conference on Lévy Processes: Theory and Applications Wroclaw, July 15 19,
More informationThe multidimensional Ito Integral and the multidimensional Ito Formula. Eric Mu ller June 1, 2015 Seminar on Stochastic Geometry and its applications
The multidimensional Ito Integral and the multidimensional Ito Formula Eric Mu ller June 1, 215 Seminar on Stochastic Geometry and its applications page 2 Seminar on Stochastic Geometry and its applications
More informationSquared Bessel Process with Delay
Southern Illinois University Carbondale OpenSIUC Articles and Preprints Department of Mathematics 216 Squared Bessel Process with Delay Harry Randolph Hughes Southern Illinois University Carbondale, hrhughes@siu.edu
More informationStrong Solutions and a Bismut-Elworthy-Li Formula for Mean-F. Formula for Mean-Field SDE s with Irregular Drift
Strong Solutions and a Bismut-Elworthy-Li Formula for Mean-Field SDE s with Irregular Drift Thilo Meyer-Brandis University of Munich joint with M. Bauer, University of Munich Conference for the 1th Anniversary
More informationSome Properties of NSFDEs
Chenggui Yuan (Swansea University) Some Properties of NSFDEs 1 / 41 Some Properties of NSFDEs Chenggui Yuan Swansea University Chenggui Yuan (Swansea University) Some Properties of NSFDEs 2 / 41 Outline
More informationContinuous Time Finance
Continuous Time Finance Lisbon 2013 Tomas Björk Stockholm School of Economics Tomas Björk, 2013 Contents Stochastic Calculus (Ch 4-5). Black-Scholes (Ch 6-7. Completeness and hedging (Ch 8-9. The martingale
More informationExercises. T 2T. e ita φ(t)dt.
Exercises. Set #. Construct an example of a sequence of probability measures P n on R which converge weakly to a probability measure P but so that the first moments m,n = xdp n do not converge to m = xdp.
More informationConvergence at first and second order of some approximations of stochastic integrals
Convergence at first and second order of some approximations of stochastic integrals Bérard Bergery Blandine, Vallois Pierre IECN, Nancy-Université, CNRS, INRIA, Boulevard des Aiguillettes B.P. 239 F-5456
More informationStudy of Dependence for Some Stochastic Processes
Study of Dependence for Some Stochastic Processes Tomasz R. Bielecki, Andrea Vidozzi, Luca Vidozzi Department of Applied Mathematics Illinois Institute of Technology Chicago, IL 6616, USA Jacek Jakubowski
More informationLecture 22 Girsanov s Theorem
Lecture 22: Girsanov s Theorem of 8 Course: Theory of Probability II Term: Spring 25 Instructor: Gordan Zitkovic Lecture 22 Girsanov s Theorem An example Consider a finite Gaussian random walk X n = n
More informationON THE FIRST TIME THAT AN ITO PROCESS HITS A BARRIER
ON THE FIRST TIME THAT AN ITO PROCESS HITS A BARRIER GERARDO HERNANDEZ-DEL-VALLE arxiv:1209.2411v1 [math.pr] 10 Sep 2012 Abstract. This work deals with first hitting time densities of Ito processes whose
More informationA NOTE ON STATE ESTIMATION FROM DOUBLY STOCHASTIC POINT PROCESS OBSERVATION. 0. Introduction
STUDIA UNIV. BABEŞ BOLYAI, MATHEMATICA, Volume XLVI, Number 1, March 1 A NOTE ON STATE ESTIMATION FROM DOUBLY STOCHASTIC POINT PROCESS OBSERVATION. Introduction In this note we study a state estimation
More informationarxiv: v2 [math.pr] 27 Oct 2015
A brief note on the Karhunen-Loève expansion Alen Alexanderian arxiv:1509.07526v2 [math.pr] 27 Oct 2015 October 28, 2015 Abstract We provide a detailed derivation of the Karhunen Loève expansion of a stochastic
More informationAn application of hidden Markov models to asset allocation problems
Finance Stochast. 1, 229 238 (1997) c Springer-Verlag 1997 An application of hidden Markov models to asset allocation problems Robert J. Elliott 1, John van der Hoek 2 1 Department of Mathematical Sciences,
More informationThe concentration of a drug in blood. Exponential decay. Different realizations. Exponential decay with noise. dc(t) dt.
The concentration of a drug in blood Exponential decay C12 concentration 2 4 6 8 1 C12 concentration 2 4 6 8 1 dc(t) dt = µc(t) C(t) = C()e µt 2 4 6 8 1 12 time in minutes 2 4 6 8 1 12 time in minutes
More informationDensities for the Navier Stokes equations with noise
Densities for the Navier Stokes equations with noise Marco Romito Università di Pisa Universitat de Barcelona March 25, 2015 Summary 1 Introduction & motivations 2 Malliavin calculus 3 Besov bounds 4 Other
More informationMultiple points of the Brownian sheet in critical dimensions
Multiple points of the Brownian sheet in critical dimensions Robert C. Dalang Ecole Polytechnique Fédérale de Lausanne Based on joint work with: Carl Mueller Multiple points of the Brownian sheet in critical
More informationStochastic Integration and Continuous Time Models
Chapter 3 Stochastic Integration and Continuous Time Models 3.1 Brownian Motion The single most important continuous time process in the construction of financial models is the Brownian motion process.
More informationONLINE APPENDIX TO: NONPARAMETRIC IDENTIFICATION OF THE MIXED HAZARD MODEL USING MARTINGALE-BASED MOMENTS
ONLINE APPENDIX TO: NONPARAMETRIC IDENTIFICATION OF THE MIXED HAZARD MODEL USING MARTINGALE-BASED MOMENTS JOHANNES RUF AND JAMES LEWIS WOLTER Appendix B. The Proofs of Theorem. and Proposition.3 The proof
More informationCOVARIANCE IDENTITIES AND MIXING OF RANDOM TRANSFORMATIONS ON THE WIENER SPACE
Communications on Stochastic Analysis Vol. 4, No. 3 (21) 299-39 Serials Publications www.serialspublications.com COVARIANCE IDENTITIES AND MIXING OF RANDOM TRANSFORMATIONS ON THE WIENER SPACE NICOLAS PRIVAULT
More informationChange of Measure formula and the Hellinger Distance of two Lévy Processes
Change of Measure formula and the Hellinger Distance of two Lévy Processes Erika Hausenblas University of Salzburg, Austria Change of Measure formula and the Hellinger Distance of two Lévy Processes p.1
More informationGeneralized Gaussian Bridges
TOMMI SOTTINEN ADIL YAZIGI Generalized Gaussian Bridges PROCEEDINGS OF THE UNIVERSITY OF VAASA WORKING PAPERS 4 MATHEMATICS 2 VAASA 212 Vaasan yliopisto University of Vaasa PL 7 P.O. Box 7 (Wolffintie
More informationStrong uniqueness for stochastic evolution equations with possibly unbounded measurable drift term
1 Strong uniqueness for stochastic evolution equations with possibly unbounded measurable drift term Enrico Priola Torino (Italy) Joint work with G. Da Prato, F. Flandoli and M. Röckner Stochastic Processes
More informationAsymptotic Properties of an Approximate Maximum Likelihood Estimator for Stochastic PDEs
Asymptotic Properties of an Approximate Maximum Likelihood Estimator for Stochastic PDEs M. Huebner S. Lototsky B.L. Rozovskii In: Yu. M. Kabanov, B. L. Rozovskii, and A. N. Shiryaev editors, Statistics
More informationThe Codimension of the Zeros of a Stable Process in Random Scenery
The Codimension of the Zeros of a Stable Process in Random Scenery Davar Khoshnevisan The University of Utah, Department of Mathematics Salt Lake City, UT 84105 0090, U.S.A. davar@math.utah.edu http://www.math.utah.edu/~davar
More informationAsymptotics for posterior hazards
Asymptotics for posterior hazards Pierpaolo De Blasi University of Turin 10th August 2007, BNR Workshop, Isaac Newton Intitute, Cambridge, UK Joint work with Giovanni Peccati (Université Paris VI) and
More informationInterest Rate Models:
1/17 Interest Rate Models: from Parametric Statistics to Infinite Dimensional Stochastic Analysis René Carmona Bendheim Center for Finance ORFE & PACM, Princeton University email: rcarmna@princeton.edu
More informationShort-time expansions for close-to-the-money options under a Lévy jump model with stochastic volatility
Short-time expansions for close-to-the-money options under a Lévy jump model with stochastic volatility José Enrique Figueroa-López 1 1 Department of Statistics Purdue University Statistics, Jump Processes,
More informationEULER MARUYAMA APPROXIMATION FOR SDES WITH JUMPS AND NON-LIPSCHITZ COEFFICIENTS
Qiao, H. Osaka J. Math. 51 (14), 47 66 EULER MARUYAMA APPROXIMATION FOR SDES WITH JUMPS AND NON-LIPSCHITZ COEFFICIENTS HUIJIE QIAO (Received May 6, 11, revised May 1, 1) Abstract In this paper we show
More informationWHITE NOISE APPROACH TO THE ITÔ FORMULA FOR THE STOCHASTIC HEAT EQUATION
Communications on Stochastic Analysis Vol. 1, No. 2 (27) 311-32 WHITE NOISE APPROACH TO THE ITÔ FORMULA FOR THE STOCHASTIC HEAT EQUATION ALBERTO LANCONELLI Abstract. We derive an Itô s-type formula for
More informationStochastic Calculus February 11, / 33
Martingale Transform M n martingale with respect to F n, n =, 1, 2,... σ n F n (σ M) n = n 1 i= σ i(m i+1 M i ) is a Martingale E[(σ M) n F n 1 ] n 1 = E[ σ i (M i+1 M i ) F n 1 ] i= n 2 = σ i (M i+1 M
More informationCONDITIONAL FULL SUPPORT OF GAUSSIAN PROCESSES WITH STATIONARY INCREMENTS
J. Appl. Prob. 48, 561 568 (2011) Printed in England Applied Probability Trust 2011 CONDITIONAL FULL SUPPOT OF GAUSSIAN POCESSES WITH STATIONAY INCEMENTS DAIO GASBAA, University of Helsinki TOMMI SOTTINEN,
More informationA New Fractional Process: A Fractional Non-homogeneous Poisson Process Enrico Scalas
A New Fractional Process: A Fractional Non-homogeneous Poisson Process Enrico Scalas University of Sussex Joint work with Nikolai Leonenko (Cardiff University) and Mailan Trinh Fractional PDEs: Theory,
More informationLimit theorems for multipower variation in the presence of jumps
Limit theorems for multipower variation in the presence of jumps Ole E. Barndorff-Nielsen Department of Mathematical Sciences, University of Aarhus, Ny Munkegade, DK-8 Aarhus C, Denmark oebn@imf.au.dk
More informationWeak solutions of mean-field stochastic differential equations
Weak solutions of mean-field stochastic differential equations Juan Li School of Mathematics and Statistics, Shandong University (Weihai), Weihai 26429, China. Email: juanli@sdu.edu.cn Based on joint works
More informationUniversal examples. Chapter The Bernoulli process
Chapter 1 Universal examples 1.1 The Bernoulli process First description: Bernoulli random variables Y i for i = 1, 2, 3,... independent with P [Y i = 1] = p and P [Y i = ] = 1 p. Second description: Binomial
More information(2m)-TH MEAN BEHAVIOR OF SOLUTIONS OF STOCHASTIC DIFFERENTIAL EQUATIONS UNDER PARAMETRIC PERTURBATIONS
(2m)-TH MEAN BEHAVIOR OF SOLUTIONS OF STOCHASTIC DIFFERENTIAL EQUATIONS UNDER PARAMETRIC PERTURBATIONS Svetlana Janković and Miljana Jovanović Faculty of Science, Department of Mathematics, University
More informationExponential martingales: uniform integrability results and applications to point processes
Exponential martingales: uniform integrability results and applications to point processes Alexander Sokol Department of Mathematical Sciences, University of Copenhagen 26 September, 2012 1 / 39 Agenda
More informationStochastic Integration.
Chapter Stochastic Integration..1 Brownian Motion as a Martingale P is the Wiener measure on (Ω, B) where Ω = C, T B is the Borel σ-field on Ω. In addition we denote by B t the σ-field generated by x(s)
More informationAn Overview of the Martingale Representation Theorem
An Overview of the Martingale Representation Theorem Nuno Azevedo CEMAPRE - ISEG - UTL nazevedo@iseg.utl.pt September 3, 21 Nuno Azevedo (CEMAPRE - ISEG - UTL) LXDS Seminar September 3, 21 1 / 25 Background
More informationConvergence of Particle Filtering Method for Nonlinear Estimation of Vortex Dynamics
Convergence of Particle Filtering Method for Nonlinear Estimation of Vortex Dynamics Meng Xu Department of Mathematics University of Wyoming February 20, 2010 Outline 1 Nonlinear Filtering Stochastic Vortex
More informationCitation Osaka Journal of Mathematics. 41(4)
TitleA non quasi-invariance of the Brown Authors Sadasue, Gaku Citation Osaka Journal of Mathematics. 414 Issue 4-1 Date Text Version publisher URL http://hdl.handle.net/1194/1174 DOI Rights Osaka University
More informationThe Lévy-Itô decomposition and the Lévy-Khintchine formula in31 themarch dual of 2014 a nuclear 1 space. / 20
The Lévy-Itô decomposition and the Lévy-Khintchine formula in the dual of a nuclear space. Christian Fonseca-Mora School of Mathematics and Statistics, University of Sheffield, UK Talk at "Stochastic Processes
More informationSolving the Poisson Disorder Problem
Advances in Finance and Stochastics: Essays in Honour of Dieter Sondermann, Springer-Verlag, 22, (295-32) Research Report No. 49, 2, Dept. Theoret. Statist. Aarhus Solving the Poisson Disorder Problem
More informationGARCH processes continuous counterparts (Part 2)
GARCH processes continuous counterparts (Part 2) Alexander Lindner Centre of Mathematical Sciences Technical University of Munich D 85747 Garching Germany lindner@ma.tum.de http://www-m1.ma.tum.de/m4/pers/lindner/
More informationA REPRESENTATION FOR THE KANTOROVICH RUBINSTEIN DISTANCE DEFINED BY THE CAMERON MARTIN NORM OF A GAUSSIAN MEASURE ON A BANACH SPACE
Theory of Stochastic Processes Vol. 21 (37), no. 2, 2016, pp. 84 90 G. V. RIABOV A REPRESENTATION FOR THE KANTOROVICH RUBINSTEIN DISTANCE DEFINED BY THE CAMERON MARTIN NORM OF A GAUSSIAN MEASURE ON A BANACH
More informationMaximum Likelihood Drift Estimation for Gaussian Process with Stationary Increments
Austrian Journal of Statistics April 27, Volume 46, 67 78. AJS http://www.ajs.or.at/ doi:.773/ajs.v46i3-4.672 Maximum Likelihood Drift Estimation for Gaussian Process with Stationary Increments Yuliya
More informationConvoluted Brownian motions: a class of remarkable Gaussian processes
Convoluted Brownian motions: a class of remarkable Gaussian processes Sylvie Roelly Random models with applications in the natural sciences Bogotá, December 11-15, 217 S. Roelly (Universität Potsdam) 1
More informationRegularity of the density for the stochastic heat equation
Regularity of the density for the stochastic heat equation Carl Mueller 1 Department of Mathematics University of Rochester Rochester, NY 15627 USA email: cmlr@math.rochester.edu David Nualart 2 Department
More informationOptimal portfolio strategies under partial information with expert opinions
1 / 35 Optimal portfolio strategies under partial information with expert opinions Ralf Wunderlich Brandenburg University of Technology Cottbus, Germany Joint work with Rüdiger Frey Research Seminar WU
More informationMA8109 Stochastic Processes in Systems Theory Autumn 2013
Norwegian University of Science and Technology Department of Mathematical Sciences MA819 Stochastic Processes in Systems Theory Autumn 213 1 MA819 Exam 23, problem 3b This is a linear equation of the form
More informationA. R. SOLTANI AND Z. SHISHEBOR
GEORGIAN MAHEMAICAL JOURNAL: Vol. 6, No. 1, 1999, 91-98 WEAKLY PERIODIC SEQUENCES OF BOUNDED LINEAR RANSFORMAIONS: A SPECRAL CHARACERIZAION A. R. SOLANI AND Z. SHISHEBOR Abstract. Let X and Y be two Hilbert
More informationLecture 2: From Linear Regression to Kalman Filter and Beyond
Lecture 2: From Linear Regression to Kalman Filter and Beyond Department of Biomedical Engineering and Computational Science Aalto University January 26, 2012 Contents 1 Batch and Recursive Estimation
More informationFinite element approximation of the stochastic heat equation with additive noise
p. 1/32 Finite element approximation of the stochastic heat equation with additive noise Stig Larsson p. 2/32 Outline Stochastic heat equation with additive noise du u dt = dw, x D, t > u =, x D, t > u()
More informationITO OLOGY. Thomas Wieting Reed College, Random Processes 2 The Ito Integral 3 Ito Processes 4 Stocastic Differential Equations
ITO OLOGY Thomas Wieting Reed College, 2000 1 Random Processes 2 The Ito Integral 3 Ito Processes 4 Stocastic Differential Equations 1 Random Processes 1 Let (, F,P) be a probability space. By definition,
More informationLecture 4: Ito s Stochastic Calculus and SDE. Seung Yeal Ha Dept of Mathematical Sciences Seoul National University
Lecture 4: Ito s Stochastic Calculus and SDE Seung Yeal Ha Dept of Mathematical Sciences Seoul National University 1 Preliminaries What is Calculus? Integral, Differentiation. Differentiation 2 Integral
More informationLévy Processes and Infinitely Divisible Measures in the Dual of afebruary Nuclear2017 Space 1 / 32
Lévy Processes and Infinitely Divisible Measures in the Dual of a Nuclear Space David Applebaum School of Mathematics and Statistics, University of Sheffield, UK Talk at "Workshop on Infinite Dimensional
More information