Lecture notes for Numerik IVc - Numerics for Stochastic Processes, Wintersemester 2012/2013. Instructor: Prof. Carsten Hartmann

Size: px
Start display at page:

Download "Lecture notes for Numerik IVc - Numerics for Stochastic Processes, Wintersemester 2012/2013. Instructor: Prof. Carsten Hartmann"

Transcription

1 Lecture notes for Numerik IVc - Numerics for Stochastic Processes, Wintersemester 22/23. Instructor: Prof. Carsten Hartmann Scribe: H. Lie Course outline. Probability theory (a) Some basics: stochastic processes, conditional probabilities and expectations, Markov chains References: [MS5, Kle6] 2. Stochastic differential equations (a) Brownian motion: properties of the paths, Strong Markov Property References: [MS5, Øks3, Arn73] (b) Stochastic integrals: Itô integrals, Itô calculus, Itô isometry References: [MS5, Øks3, Arn73] (c) SDEs: existence and uniqueness of solutions, numerical discretisation, applications from physics, biolopgy and finance References: [Øks3, Arn73, KP92] (d) Misc: Kolmogorov forward and backward PDEs, infinitesimal generators, semigroup theory, stopping times, invariant distributions, Markov Chain Monte Carlo methods for PDEs and SDEs References: [Øks3, Arn73, KP92] 3. Filtering theory (if time permits) (a) Linear filtering: conditional expectation, best approximation, Kalman- Bucy filter for SDEs Reference: [Jaz7, Øks3] 4. Approximation of stochastic processes (a) Spectral theory of Markov chains: infinitesimal generator, metastability, aggregation of Markov chains References: [HM5, Sar] (b) Markov jump processes: applications from biology, physiocs and finance, Markov decision processes, control theory References: [GHL9]

2 Figure : Simulation of a butane molecule and its approximation by a 3-state Markov chain (states in blue, green and yellow; solvent molecules not shown). Day, Different levels of modelling.. Time-discrete Markov chains Time index set I is discrete, e.g. I N and state space S is countable or finite, e.g. S = {s, s 2, s 3 } (see Figure ). Key objects are transition probabilities. For a state space S = {,..., n}, the transition probabilities p ij satisfy p ij = P (X t+ = j X t = i) and yield a row-stochastic matrix P = (p ij ) i,j S...2 Markov jump processes These are time-continuous, discrete state-space Markov chains. Time index set I R +, S discrete. For a fixed time step h >, the transition probabilities are given by (see Figure 2) P (X t+h = s j X t = s i ) = hl ij + o(h) where L = (l ij ) i,j S and P h are matrices satisfying P h = exp (hl). Note: the matrix L is row sum zero, i.e. j l ij =. The waiting times for the Markov chain in any state s i are exponentially distributed in the sense that P (X t+s = s i, s [, τ) X t = s i ) = exp (l ii τ) 2

3 Figure 2: Simulation of butane: typical time series of the central dihedral angle (blue: metastable diffusion process, red: Markov jump process) and the average waiting time is l ii (by definition of the exponential distribution). Note: the spectrum of the matrix P h is contained within the unit disk, i.e. for every eigenvalue λ of P h, λ. This property is a consequence of P h being row-stochastic, i.e. that j P h,ij =. Since P h = exp(hl) it follows that σ(p h ) D := { x R 2 x } σ(l) C = {y C Re(y) } Example.. Suppose one has a reversible reaction in which one has a large collection of N molecules of the same substance. The molecules can be either in state A or state B and the molecules can change between the two states. Let k + denote the rate of the reaction in which molecules change from state A to B and let k denote the rate at which molecules change from state B to A. For t >, consider the quantity µ A i (t) := P (number of molecules in state A at time t is i) where i = {,..., N}. One can define quantities µ B i (t) in a similar way, and one can construct balance laws for these quantities, e.g. dµ A i (t) dt = k + µ A i+(t) + k µ A i (t) (k + + k )µ A i (t). The above balance law can be written in vector notation using a tridiagonal matrix L. By adding an initial condition one can obtain an initial value problem dµ A (t) dt = L µ A (t), µ A () = µ. The solution of the initial value problem above is µ A (t) = µ exp ( tl ). 3

4 ..3 Stochastic differential equations (SDEs) These are time-continuous, continuous state space Markov chains. SDEs may be considered to be ordinary differential equations (ODEs) with an additional noise term (cf. Figure 2). Let b : R n R n be a smooth vector field and let x(t) be a deterministic dynamical system governed by the vector field b( ). Then x(t) evolves according to dx dt = b(x), x() = x. () Now let (B t ) t> be Brownian motion in R d, and let (X t ) t> be a dynamical system in R d which evolves according to the equation dx t dt = b(x t ) + db t dt. (2) The additional term dbt dt represents noise, or random perturbations from the environment, but is not well-defined because the paths of Brownian motion are nowhere differentiable. Therefore, one sometimes writes which is shorthand for X t = X + dx t = b(x t )dt + db t, t b(x t )dt + t db t. The most common numerical integration method for SDEs is the forward Euler method. If x is a C function of time t, then dx x(s + h) x(s) dt = lim. t=s h h The forward Euler method for ODEs of the form () is given by X t+h = X t + hb(x t ) and for SDEs of the form (2) it is given by X t+h = X t + hb(x t ) + ξ h where < h is the integration time step and the noise term ξ in the Euler method for SDEs is modeled by a mean-zero Gaussian random variable. For stochastic dynamical systems which evolve according to SDEs as in (2), one can consider the probability that a system at some point x R d will be in a set A R d after a short time h > : P (X t+h A X t = x). The associated transition probability density functions of these stochastic dynamical systems are Gaussian because the noise term in (2) is Gaussian. What has been the generator matrix L in case of a Markov jump process is an infinite-dimensional operator acting on a suitable Banach space. Specifically, E x [f(x t )] f(x ) Lf(x ) = lim, t t provided that the limit exists. Here f : R n R is any measurable function and E x [ ] denotes the expectation over all random paths of X t satisfying X = x. L is a second-order differential operator if f is twice differentiable. 4

5 2 Day 2, Preliminaries from probability theory Let (Ω, E, P) be a probability space, where Ω is a set and E 2 Ω is a σ-field or σ-algebra on Ω, and P is a probability measure (i.e., P is a nonnegative, countably additive measure on (Ω, E) with the property P(Ω) = ). 2. Conditioning Let A E be a set of nonzero measure, i.e. P(A) > and define E A to be the set of all subsets of A which are elements of E, i.e. E A := {E A E E}. Definition 2. (Conditional probability, part I). For an event A and an event E E A, the conditional probability of E given A is P(E A) := P(E A). P(A) Remark 2.2. Think of P A := P( A) as a probability measure on the measurable space (A, E A ). Given a set B E, the characteristic or indicator function χ B : Ω {, } satisfies { x B χ B (x) = x / B. Definition 2.3 (Conditional expectation, part I). Let X : Ω R be a random variable with finite expectation with respect to P. The conditional expectation of X given an event A is Remark 2.4. We have E(X A) = E [Xχ A] P(A) E(X A) = P(A) Remark 2.5. Observe that P(E A) = E [χ E A]. A. XdP = XdP A. Up to this point we have only considered the case where A satisfies P(A) >. We now consider the general case. Definition 2.6 (Conditional expectation, part II). Let X : Ω R be an integrable random variable with respect to P and let F E be any sub-sigma algebra of E. The conditional expectation of X given F is a random variable Y := E [X F] with the following properties: Y is measurable with respect to F: B B(R), Y (B) F. We have XdP = Y dp F F. F F 5

6 Remark 2.7. The second condition in the last definition amounts to the projection property as can be seen by noting that E [Xχ F ] = XdP = Y dp = E [Y χ F ] = E [E [X F] χ F ]. F F By the Radon-Nikodym theorem [MS5], the conditional expectation exists and is unique up to P-null sets. Definition 2.8 (Conditional probability, part II). Define the conditional probability of an event E E given A by P(E A) := E [χ E A] Exercise 2.9. Let X, Y : Ω R and scalars a, b R. Prove the following properties of the conditional expectation: (Linearity): (Law of total expectation): (Law of total probability): E [ax + by A] = ae [X A] + be [Y A]. E [X] = E [X A] + P(A) + E [X A c ] P(A c ) P(B) = P(B A)P(A) + P(B A c )P(A c ). Example 2.. The following is a collection of standard examples. Gaussian random variables: Let X, X 2 be jointly Gaussian with distribution N(µ, Σ), where ( ) ( ) E[X ] a b µ =, Σ = E[X 2 ] b c such that Σ is positive definite. The density of the distribution is [ ρ(x) = exp ] det(2πσ) 2 (x µ) Σ (x µ) (Ex.: Compute the distribution of X given that X 2 = a for some a R.) (Conditioning as coarse-graining): Let Z = {Z i } M i= i.e. Ω = M i= Z i with Z i Z j = and define be a partition of Ω, Y (ω) = M E [X Z i ] χ Zi (ω). Then Y = E [X Z] is a conditional expectation (cf. Figure 3) i= (Exponential waiting times): exponential waiting times are random variables T : Ω [, ) with the memoryless property: P (T > s + t T > s) = P (T > t). This property is equivalent to the statement that T has an exponential distribution, i.e. that P(T > t) = exp( λt) for a parameter value λ >. 6

7 Figure 3: Simulation of butane, coarse-grained into three states Z, Z 2, Z Stochastic processes Definition 2. (Stochastic process). A stochastic process X = {X t } t I is a collection of random variables on a probability space (Ω, E, P) indexed by a parameter t I [, ). We call X discrete in time if I N continuous in time if I = [, T ] for any T <. How does one define probabilities for X? We provide a basic argument to illustrate the possible difficulties in defining the probability of a stochastic process in an unambiguous way. By definition of a stochastic process, X t = X t (ω) is measurable for every fixed t I, but if one has an event of the form E = {ω Ω X t (ω) [a, b] t I} how does one define the probability of this event? If t is discrete, the σ-additivity of P saves us, together with the measurability of X t for every t. If, however, the process is time-continuous, X t is defined only almost surely (a.s.) and we are free to change X t on a set A t with P(A t ) =. By this method we can change X t on A = t I A t, The problem now is that P(A) need not be equal to zero even though P(A t ) = t I. Furthermore, P(E) may not be uniquely defined. So what can we do? The solution to the question of how to define probabilities for stochastic processes is to use finite-dimensional distributions or marginals. Definition 2.2. (Finite dimensional distributions): Fix d N, t,..., t d I. The finite-dimensional distributions of the stochastic process X for (t,..., t d ) are defined as µ t,...,t d (B) := P (Xtk ) k=,...,d (B) = P ({ω Ω (X t (ω),..., X td (ω)) B}) for B B(R d ). 7

8 Here and in the following we use the shorthand notation P Y := P Y to denote the push forward of P by the random variable Y. Theorem 2.3. (Kolmogorov Extension Theorem): Fix d N, t,..., t d I, and let µ t,...,t d be a consistent family of finite-dimensional distributions, i.e. for any permutation π of (,..., d), µ t,...,t d (B... B d ) = µ (tπ(),...,t π(d) (B π()... B π(d) ) For t,..., t d+ I, we have that µ t,...,t d+ (B... B d R) = µ t,...,t d (B... B d ). Then there exists a stochastic process X = (X t ) t I with µ t,...t d as its finitedimensional distribution. Remark 2.4. The Kolmogorov Extension Theorem does not guarantee uniqueness, not even P-a.s. uniqueness, and, as we will see later on, such a kind of uniqueness would not be a desirable property of a stochastic process. Definition 2.5. (Filtration generated by a stochastic process X): Let F = {F t } t I with F s F t for s < t be a filtration generated by F t = σ ({X s s t}) is called the filtration generated by X. 2.3 Markov processes Definition 2.6. A stochastic process X is a Markov process if where for some event E. P (X t+s A F s ) = P (X t+s A X s ) (3) P ( X s ) := P ( σ(x s )), P (E σ(x s )) := E [χ E σ(x s )] Remark 2.7. If I is discrete, then X is a Markov process if P (X n+ A X = x,..., X n = x n ) = P (X n+ A X n = x n ) Example 2.8. Consider a Markov Chain (X t ) t N on a continuous state space S R and let S be a σ-algebra on S. Let the evolution of (X t ) t N be described by the transition kernel p(, ) : S S [, ] which gives the single-step transition probabilities: p(x, A) := P (X t+ A X t = x) = q(x, y)dy. A In the above, A B(S) and q = dp dλ is the density of the transition kernel with respect to Lebesgue measure. The transition kernel has the property that 8

9 x S, p(x, ) is a probability measure on S, while for every A S, p(, A) is a measurable function on S. For a concrete example, consider the Euler-Maruyama discretization of an SDE for a fixed time step t, X n+ = X n + tξ n+, X =, where (ξ i ) i N are independent, identically distributed (i.i.d) Gaussian N (, ) random variables. The process (X i ) i N is a Markov Chain on R. The transition kernel p(x, A) has the Gaussian transition density q(x, y) = 2π t exp [ 2 y x 2 Thus, if X n = x, then the probability that X n+ A R is given by P (X n+ A X n = x) = q(x, y)dy. 3 Day 3, Recapitulation: A stochastic process X = (X t ) t I is a collection of random variables X t : Ω R indexed by t I (e.g. I = [, )) on some probability space (Ω, E, P). A filtration F := (F t ) t I is a collection of increasing sigma-algebras satisfying F t F s for t < s. A stochastic process X is said to be adapted to F if (X s ) s t is F t -measurable. For example, if we define F t := σ(x s : s t), then X is adapted to F. The probability distribution of a random variable X is given in terms of its finite dimensional distributions. Example 3. (Continued from last week). Let I = N and consider a sequence (X n ) n N of random variables X n = Xn t governed by the relation A X t n+ = X t n + tξ n+, X t = a.s. (4) where t >, and (ξ k ) k N are i.i.d random variables with E [ξ k ] = and E [ ξk] 2 = (not necessarily Gaussian). To obtain a continuous-time stochastic process, the values of the stochastic process on non-integer time values may be obtained by linear interpolation (cf. Figure 4 below). We want to consider the limiting behaviour of the stochastic process in the limit as t goes to zero. Set t = t/n for a fixed terminal time t < and let N ( t ). Then, by the central limit theorem, t N XN t = ξ k tz (5) N k= where Z N (, ), and means convergence in distribution, i.e., weak convergence of the induced probability measure; equivalently, the limiting random variable is distributed according to N (, t). In other words the limiting t ]. 9

10 distribution of the random variable X t N for fixed t = N t is the same as the distribution of a centered Gaussian random variable with variance t. As this is true for any t >, we can think of the limiting process as a continuous-time Markov process B = (B t ) t> with Gaussian transition probabilities, P (B t+s A B s = x) = = A q s,t (x, y)dy 2π t s A exp ( y x 2 2 t s ) dy. The stochastic process B is homogeneous or time-homogeneous because the transition probability density q s,t (, ) does not depend on the actual values of t and s, but only on their difference, i.e., q s,t (, ) = q s t (, ) (6) X t X t t t 2 x 4.4 X t t 5 5 x Figure 4: Sample paths of (Xn t ) n for t =.5,.2,. over the unit time interval [, ], with piecewise constant interpolation. The lower right plot shows the histogram (i.e., the unnormalized empirical distribution) of (X) t at time t =, averaged over independent realizations. Remark 3.2. The choice of exponent /2 in t = ( t) /2 in (5) is unique. For ( t) α with α (, 2 ), the limit of X t n explodes in the sense that the variance of the process blows up, i.e., E[(XN t)2 ] as N. On the other hand, for ( t) α with α > /2, XN t in probability as N.

11 3. Brownian motion Brownian motion is named after the British botanist, Robert Brown ( ), who first observed the random motion of pollen particles suspended in water. Einstein called the Brownian process Zitterbewegung in his 95 paper, Über die von der molekularkinetischen Theorie der Wärme gefordete Bewegung von in ruhenden Flüssigkeiten suspendierten Teilchen. The Brownian motion is a continuous-time stochastic process which is nowhere differentiable. It is also a martingale in the sense that on average, the particle stays in the same location at which it was first observed. In other words, the best estimate of where the particle will be after a time t > is its initial location. Definition 3.3. (Brownian motion) The stochastic process B = (B t ) t> with B t R is called the -dimensional Brownian motion or the -dimensional Wiener process if it has the following properties: (i) B = P-a.s. (ii) B has independent increments, i.e., for all s < t, (B t B s ) is a random variable which is independent of B r for r s. (iii) B has stationary, Gaussian increments, i.e., for t > s we have B t B s D = Bt s (7a) D = N (, t s). (7b) (iv) Trajectories of Brownian motion are continuous functions of time. We now make precise some important notions: Definition 3.4. (Filtered probability space) A filtered probability space is a probability space (Ω, F, P) with a filtration (F t ) t such that t, F t F. Remark 3.5. One may write (Ω, F, F t, P) to refer to a filtered probability space. However, if one is working with a particular stochastic process X, one may consider the sigma-algebra F on Ω to simply be the smallest sigma-algebra which contains the union of the F X t, where F X t := σ(x s : s t). In symbols, we define the sigma-algebra in the probability space to be F := t F t := σ ( t F t ). Definition 3.6. (Martingale) A stochastic process X = (X t ) t> is a martingale with respect to a filtered probability space (Ω, F, F t, P) if X satisfies the following properties: (i) X is adapted to F, i.e. X t is measurable with respect to F t for every t (ii) X is integrable: X L (Ω, F, P), i.e. E [ X ] = X(ω) dp(ω) < The notation X D = Y means X has the same distribution as Y. Ω

12 (iii) X has the martingale property: t > s E [X t F s ] = X s. Definition 3.7. (Gaussian process) A -dimensional process G = (G t ) t> is called a Gaussian process if for any collection (t,..., t m ) I for arbitrary m N, the random variable (G t,..., G tm ) has a Gaussian distribution, i.e. it has a density [ f(g) = exp ] det(2πσ) 2 (g µ) Σ (g µ) (8) where g = (g,..., g m ), µ R m is a constant vector of means and Σ = Σ R m m is a symmetric positive semi-definite matrix. Remark 3.8. The Brownian motion process is a Gaussian process with the vector of means µ = and covariance matrix t.... Σ = t 2 t (9)..... t m t m The covariance matrix is diagonal due to the independence of the increments of Brownian motion. Remark 3.9. Some further remarks are in order. (a) Conditions (i)-(iii) define a consistent family of finite-dimensional distributions. Hence, the existence of the process B is guaranteed by the Kolmogorov Extension Theorem. (b) Conditions (i)-(iii) imply that E [B t ] = and E [B t B s ] = min(t, s) s, t R. The proof is left as an exercise. ) t> (which is obtained by linear interpolation between the XN t ) and B as random variables on the space of continuous trajectories (C(R + ) and B(C(R + ))), then the process (Xt t ) t> converges in distribution to B. (c) The discrete process (Xn t ) n N converges in distribution to a Brownian motion (B t ) t if the time discrete is linearly interpolated between two successive points. In other words, if we consider the continuous-time stochastic processes (X t t (d) We have that E [ (B t B s ) 2] = E [ (B t s ) 2] by (7a) in Definition 3.3 = t s by (7b) in Definition 3.3. (e) Brownian motion enjoys the following scaling invariance, also known as self-similarity of Brownian motion: for every t > and α >, B t D = α /2 B αt. 2

13 An alternative construction of Brownian motion Observe that we have constructed Brownian motion by starting with the scaled random walk process and using the Kolmogorov Extension Theorem. Now we present an alternative method for constructing Brownian motion that is useful for numerics, called the Karhunen-Loève expansion of Brownian motion. We will consider this expansion for Brownian motion on the unit time interval [, ]. Let {η k } k N be a collection of independent, identically distributed (i.i.d) Gaussian random variables distributed according to N (, ), and let {φ k (t)} k N be an orthonormal basis of L 2 ([, ]) = { u : [, ] R : By construction, the basis functions satisfy φ i, φ j = φ i (t)φ j (t)dt = δ ij, and we can represent any function f L 2 ([, ]) by f(t) = k N α k φ k (t) } u(t) 2 dt <. () for α k = f, φ k. We have the following result. Theorem 3.. (Karhunen-Loève): The process (W t ) t defined by is a Brownian motion. W t = k N t η k φ k (s)ds () Proof. We give only a sketch of the proof. For details, see the Appendix in [MS5], or [KS9]). The key components of the proof are to show the following: (i) The infinte sum which defines the Karhunen-Loève expansion is absolutely convergent, uniformly on [, ]. (ii) It holds that E [W t ] = and E [W t W s ] = min(s, t). 4 Day 4, Brownian motion From last week, we saw that the Brownian motion (B t ) t is a continuous-time stochastic process on R with stationary, independent, Gaussian increments a.s. continuous paths. That is, for fixed ω, each (B t ) t (ω) is a continuous trajectory in R. 3

14 Moreover the scaled random walk defined by X t n+ = X t n + tξ n+ with linear interpolation converges weakly (i.e. converges in distribution) to the Brownian motion process. Above, the (ξ n ) n N are independent, identically distributed (i.i.d) normalized Gaussian random variables (i.e. ξ n is Gaussian with mean zero and variance ). Remark 4.. Two remarks are in order. Continuity can be understood using the Lévy construction of Brownian motion on the set of dyadic rationals, D := { } k D n, D n := : k =...., 2n. 2n n N The construction of Brownian motion on the unit time interval is as follows. Let {Z t } t D be a collection of independent, normalized random variables defined on a probability space. Define the collection of functions (F n ) n N, where F n : [, ] R are given by t D n F n (t) := 2 (j+)/2 Z t t D j \ D j lin. interp. in between. Then the process B(t) = F n (t). n= is indeed a Brownian motion on [, ]. The Gaussianity of the {Z t } t D leads to the stationary, independent Gaussian increments of the process (B t ) t [,]. The continuity of the process follows from an application of the Borel-Cantelli Lemma, which states that there exists a random and almost surely finite number N N such that for all n N and d D n, Z d < c n holds. This boundedness condition implies that n N we have a decay condition for the F n : F n < c n2 n/2. Therefore the sum j F j( ) converges uniformly on [, ]. As each F j is continuous and the uniform limit of continuous functions is continuous, the process (B t ) t [,] is continuous. For more details, see [MP]. The Hausdorff dimension dim H of Brownian motion paths depends on the dimension of the space R d in which the Brownian motion paths live. 2 Let B [,] = {B t R d : t [, ]} be the graph of B t over I = [, ]. Then { 3/2 d = dim H B [,] = 2 d 2. 2 If you do not know what this is, just think of the box counting dimension that is an upper limit of the Hausdorff dimension. 4

15 The significance of this is as follows: if you consider Brownian motion paths confined to a smooth and compact two-dimensional domain and impose reflecting boundary conditions, then the Brownian motion paths will fill the domain in the limit as t. 4.2 Brownian bridge (Karhunen-Loève expansion of Brownian motion) Theorem 4.2. Let {η k } k N be i.i.d normalized random variables and {φ k } k N form a real orthonormal basis of L 2 ([, ]). Then W t = k N t η k φ k (s)ds is a Brownian motion on the interval I = [, ]. Exercise 4.3. Show that, for the definition of (W t ) t [,] above, it holds that E [W t W s ] = min(s, t) W t.2 W t t t W t t x Figure 5: Sample paths of the Karhunen-Loève expansion of (W t for M = 2, 64, 248 basis functions (you can guess which one is which). The lower right plot shows the unnormalized histogram of W t at time t =, using M = 64 basis functions and averaged over independent realizations. Remark 4.4. Unlike the scaled random walk construction of Brownian motion, no forward iterations are required here. This helps for the consideration of round-off errors in the construction of (W t ) t [,]. Furthermore: 5

16 Standard choices for the orthonormal basis {φ k } k N are Haar wavelets or trigonometric functions. Hence the numerical error can be controlled by truncating the series and by the choice of the basis. To obtain a Brownian motion on any general time interval [, T ], it suffices to use the scaling property, e.g. W [,T ] D = T W[,]/T = T k N t/t η k φ k (s)ds. 4.3 Application: filtering of Brownian motion Suppose we know that W = and W is equal to some constant ω. Without loss of generality, let ω =. Suppose we wanted to generate a Brownian motion path which interpolated between the values W = and W =. Definition 4.5. A continuous, mean-zero Gaussian process (BB t ) t is called a Brownian bridge to ω if it has the same distribution as (W t ) t [,] conditional on the terminal value W = ω. Equivalently, (BB t ) t is a Brownian bridge if Cov [BB t BB s ] = min(s, t) st. Lemma 4.6. If (W t ) t [,] is a Brownian motion, then BB t = W t tw is a Brownian bridge. Proof. Observe that E [BB t ] = E [W t tw ] = t =, so that (BB t ) t [,] is indeed mean-zero. The process (BB t ) t [,] inherits continuity from the process (W t ) t [,]. The covariance process is given by Cov(BB t BB s ) = E [BB t BB s ] = E [(W t tw ) (W s sw )] = E [W t W s ] t E [W W s ] s E [W W t ] +tse [W W ] }{{}}{{} =min(s,) =min(t,) = min(t, s) ts st + ts How does one simulate a Brownian bridge? First approach: forward iteration, using Euler s method. The time interval is [, ] and we have a time step of t := /N, so we have (N + ) discretized time nodes (t n = n t) n=,...,n and (N +) values (Yn t ) n=,...,n. Let {ξ n } n=,...,n be a collection of i.i.d normalized random variables. Forward iteration gives ( Yn+ t = Yn t t ) + tξ n+. t n It holds that t N = t by definition of t = /N. Therefore from the formula above we have YN t = tξ N+. 6

17 Therefore YN t is a mean zero Gaussian random variable with variance t. While this implies that YN t should converge in probability to the value as the step size t, the forward iteration approach is not optimal because the random variable ξ N+ is continuous, so P ( Y t N = ) =. Therefore this construction of the Brownian bridge to the value ω = will in general not yield processes which are at at time t =. As a matter of fact, YN t is unbounded and can be arbitrarily far away from zero. Second approach: Recall the Karhunen-Loéve construction of Brownian motion and choose trigonometric functions as an orthonormal basis. Then the process (W t ) t [,] given by W t (ω) = 2 M k= η k (ω) sin((k 2 )πt (k 2 )π is a Brownian motion and we can define the Brownian bridge to ω at t = by Remark 4.7. It holds that BB t = W t t(w ω). BB t = 2 k N η k sin(kπt) kπ = k N η k λk ψ k (t), where {λ k, ψ k } k N = { 2/kπ, sin(kπt) } is the eigensystem of the covariance k N operator T : L ([, ]) L ([, ]) of the process (BB t ) t [,], defined by i.e., (T u)(t) = Cov(BB t BB s ) u(s)ds, }{{} =min(t,s) st T ψ k ( ) = λ k ψ k ( ). The second approach works for any stochastic process which has finite variance over a finite time interval. For details, see [Xiu]. 7

18 .4.6 Y n BB 2 t t n t Figure 6: Sample paths of the Brownian bridge approximation, using the Euler scheme with t =.5 (left panel) and Karhunen-Loève expansion with M = 2 basis functions (right panel). 5 Day 5, (Lecturer: Stefanie W.) 5. Stochastic Integration (Itô integral) Recall that Brownian motion (B t ) t> is a stochastic process with the following properties: B = P-a.s. t < t < t 2 <... < t n, the increments B ti B ti are independent for i =,..., n and Gaussian with mean and variance t i t i. t B t (ω) is continuous P-a.s. but is P-a.s. nowhere differentiable. One of the motivations for the development of the stochastic integral lies in financial mathematics, where one wishes to determine the price of an asset that evolves randomly. The French mathematician Louis Bachelier is generally considered one of the first people to model random asset prices. In his PhD thesis, Bachelier considered the following problem. Let the value S t of an asset at time t > be modeled by S t = σb t where σ > is a scalar that describes the volatility of the stock price. Let f(t) be the amount of money an individual invests in the asset in some infinitesimal time interval [t, t + dt]. Then the wealth of the individual at the end of a time interval [, T ] is given by T T f(t)ds t = σ f(t)db t. However, it is not clear what the expression db t means. In this section, we will consider what an integral with respect to db t means, and we will also consider the case when the function f depends not only on time but on the random element ω. 8

19 The first idea is to rewrite f(t)db t = f(t) db t dt dt but as Brownian motion is almost surely nowhere differentiable, we cannot write db t dt Ṫhe second idea is to proceed as in the definition of the Lebesgue integral: start with simple step functions and later extend the definition to more general functions by the Itô Isometry. Step : Consider simple functions f(t) = n a i χ (ti,t i+](t) i= where χ A is the indicator function of a set A satisfying { x A χ A (x) = x / A. Observe that f takes a finite number n of values. By the theory of Lebesgue integration, we know that the set of these simple functions is dense in L 2 ([, )). We also know that the usual Riemann integral of such a function f corresponds to the area under the graph of f, with f(t)dt = i a i (t i+ t i ) Step 2: We now extend the method above to stochastic integral with respect to Brownian motion: f(t)db t = a i (B ti+ B ti ). Remark 5.. By the equation above, it follows that the integral f(t)db t is a random variable, since the B ti are random variables. Since increments of Brownian motion are independent and Gaussian, the integral f(t)db t is normally distributed with zero mean. What about its variance? Lemma 5.2. (Itô Isometry for simple functions) For a simple function f(t) = i a iχ (ti,t i+](t), it holds that [ ( ) ] 2 E f(t)db t = (f(t)) 2 dt. 9

20 Proof: ( ) ( ( ) ) var f(t)db t = var a i Bti+ B ti i n = a 2 i var ( ) B ti+ B ti = = = i= n a 2 i (t i+ t i ) i= n a 2 i χ (ti,t i+]dt i= n i= a 2 i χ (ti,t i+]dt = (f(t)) 2 dt. On the other hand it holds ( ) [ ( ) ] 2 [ ] var f(t)db t = E f(t)db t E f(t)db t }{{} = [ ( ) ] 2 f(t)db t = E 2 Step 3: Now we extend the definition of the integral to L 2 ([, )). result is the following The main Theorem 5.3. (Itô integral for L 2 ([, )) functions) The definition of the Itô integral can be extended to elements f L 2 ([, )) by setting f(t)db t := lim n f n (t)db t where the sequence (f n ) n N is a sequence of a simple functions satisfying f n f in L 2 ([, )), i.e. ( f n f L 2 ([, )) = (f n f) 2 (t)dt) /2 n. Proof: By the Itô isometry, we can show that ( f n (t)db t ) n N is a Cauchy sequence in the L 2 space } L 2 (Ω, F, P) := {F : Ω R : F (B(R)) F, F 2 L 2 (Ω,F,P) < where F 2 L 2 (Ω,F,P) := F 2 (ω)dp(ω). 2

21 To show that the sequence ( f n (t)db t ) n N is a Cauchy sequence, let (f l ) l N be a sequence of functions converging to f in L 2 ([, )) and consider for m, n N f n (t)db t f m (t)db t L2 (Ω,F,P) ( [ ( ) ]) 2 /2 = E f n (t)db t f m (t)db t ( [ ( ) ]) 2 /2 = E f n (t) f m (t)db t ( = (f n (t) f m (t)) 2 dt) /2 (Itô isometry) = f n f m L 2 ([, )) f n f L 2 ([, )) + f m f L 2 ([, )) and using that f n f L 2 ([, )) and f m f L 2 ([, )) as m, n, the result follows. We use the fact that L 2 (Ω, F, P) is complete (see...) to see that the limit exists and is in the same space. Moreover, by the Itô isometry the limit is independent of the sequence (f n ) n N used to approximate f in L 2 ([, )). Example 5.4. Consider the random variable exp( t)db t. How is it distributed? Using the Itô Isometry, the random variable is Gaussian with mean zero and variance 2 = exp( 2t)dt. Corollary 5.5. The Itô Isometry holds as well for f L 2 ([, )), not just simple functions. Step 4 : Now we consider functions f which depend both on the random element ω as well as time t. That is, we consider stochastic integrals of stochastic processes f : [, ) Ω R with the following properties: (i) f is B F-measurable, where B is the Borel sigma-algebra on [, ) and F is a given sigma-algebra on Ω. (ii) f(t, ω) is adapted with respect to F t, where F t := σ (B s : s t) (iii) E [ f(t, ω) 2 dt ] <. Consider simple stochastic processes of the form f(t, ω) = n a i (ω)χ (ti,t i+](t). i= Then n f(t, ω)db t = a i (ω) ( ) B ti+ B ti. i= 2

22 Example 5.6. Fix n N, fix a time step t := 2 n and define the time nodes t i := i t for i =,, 2,.... Let (B t ) t> be the standard Brownian motion. Define the following processes on [, ): f (t, ω) = i N B ti (ω)χ [ti,t i+)(t) f 2 (t, ω) = i N B ti+ (ω)χ [ti,t i+)(t). Now fix T > and N such that T = t N = N t = N2 n and compute the expected values of the integrals of f and f 2 over [, T ]. By the independent increments property of Brownian motion (or the martingale property of Brownian motion), we have [ ] T E f (t, ω)db t = N i= E [ B ti ( Bti+ B ti )] =. Using the fact above with linearity of expectation, we also have [ ] T N E f 2 (t, ω)db t = E [ ( )] B ti+ Bti+ B ti = = = i= N i= N i= N i= ( [ ( )] [ ( )]) E Bti+ Bti+ B ti E Bti Bti+ B ti [ ) ] 2 E (Bti+ B ti t i+ t i = T. In the case of Riemann integration of deterministic integrals, letting n would lead to the result that both integrals above are equal. We see that for stochastic integration, this is not the case; even if we let n, the expectations of the Itô integrals would not be equal. This is because the choice of endpoint of the interval matters in stochastic integration. Choosing the left endpoint (i.e. choosing B ti ) for f and the right endpoint (i.e. B ti+ for f 2 leads to different expectations. Note also that taking the right endpoint in f 2 leads to f 2 not being adapted, since B ti+ is not measurable with respect to F t for t < t i+. Therefore, by property (ii) above, we may not integrate f 2 with respect to db t in the way we have just described. References [Arn73] Ludwig Arnold. Stochastische Differentialgleichungen: Theorie und Anwendung. Wiley & Sons, 973. [GHL9] Xianping Guo and Onésimo Hernández-Lerma. Continuous-Time Markov Decision Processes: Theory and Applications. Springer, Heidelberg,

23 [HM5] Wilhelm Huisinga and Eike Meerbach. Markov chains for everybody. Lecture Notes, Fachbereich Mathematik und Informatik, Freie Universität Berlin, 25. [Jaz7] Andrew H. Jazwinski. Stochastic Processes and Filtering Theory. Dover, New York, 27. [Kle6] Achim Klenke. Wahrscheinlichkeitstheorie. Springer, Berlin, 26. [KP92] [KS9] Peter E. Kloeden and Eckhard Platen. Numerical Solution of Stochastic Differential Equations. Springer, Berlin, 992. Ioannis Karatzas and Steven Shreve. Brownian Motion and Stochastic Calculus. Springer, New York, 99. [MP] Peter Morters and Yuval Peres. Brownian Motion (Cambridge Series in Statistical and Probabilistic Mechanics). Cambridge University Press, 2. [MS5] David Meintrup and Stefan Schäffler. Stochastik. Springer, Berlin, 25. [Øks3] [Sar] Bernt K. Øksendal. Stochastic Differential Equations: An Introduction With Applications. Springer, Berlin, 23. Marco Sarich. Multiscale Stochastic Processes. Lecture Notes, Fachbereich Mathematik und Informatik, Freie Universität Berlin, 2. [Xiu] Dongbin Xiu. Numerical Methods for Stochastic Computations: A Spectral Method Approach. Princeton University Press, 2. 23

Lecture notes for Numerik IVc - Numerics for Stochastic Processes, Wintersemester 2012/2013. Instructor: Prof. Carsten Hartmann

Lecture notes for Numerik IVc - Numerics for Stochastic Processes, Wintersemester 2012/2013. Instructor: Prof. Carsten Hartmann Lecture notes for Numerik IVc - Numerics for Stochastic Processes, Wintersemester 212/213. Instructor: Prof. Carsten Hartmann Scribe: H. Lie December 15, 215 (updated version) Contents 1 Day 1, 16.1.212:

More information

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3 Brownian Motion Contents 1 Definition 2 1.1 Brownian Motion................................. 2 1.2 Wiener measure.................................. 3 2 Construction 4 2.1 Gaussian process.................................

More information

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS PROBABILITY: LIMIT THEOREMS II, SPRING 218. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please

More information

Lecture 4: Introduction to stochastic processes and stochastic calculus

Lecture 4: Introduction to stochastic processes and stochastic calculus Lecture 4: Introduction to stochastic processes and stochastic calculus Cédric Archambeau Centre for Computational Statistics and Machine Learning Department of Computer Science University College London

More information

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS PROBABILITY: LIMIT THEOREMS II, SPRING 15. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please

More information

ELEMENTS OF PROBABILITY THEORY

ELEMENTS OF PROBABILITY THEORY ELEMENTS OF PROBABILITY THEORY Elements of Probability Theory A collection of subsets of a set Ω is called a σ algebra if it contains Ω and is closed under the operations of taking complements and countable

More information

Brownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539

Brownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539 Brownian motion Samy Tindel Purdue University Probability Theory 2 - MA 539 Mostly taken from Brownian Motion and Stochastic Calculus by I. Karatzas and S. Shreve Samy T. Brownian motion Probability Theory

More information

Lecture 9. d N(0, 1). Now we fix n and think of a SRW on [0,1]. We take the k th step at time k n. and our increments are ± 1

Lecture 9. d N(0, 1). Now we fix n and think of a SRW on [0,1]. We take the k th step at time k n. and our increments are ± 1 Random Walks and Brownian Motion Tel Aviv University Spring 011 Lecture date: May 0, 011 Lecture 9 Instructor: Ron Peled Scribe: Jonathan Hermon In today s lecture we present the Brownian motion (BM).

More information

I forgot to mention last time: in the Ito formula for two standard processes, putting

I forgot to mention last time: in the Ito formula for two standard processes, putting I forgot to mention last time: in the Ito formula for two standard processes, putting dx t = a t dt + b t db t dy t = α t dt + β t db t, and taking f(x, y = xy, one has f x = y, f y = x, and f xx = f yy

More information

A Concise Course on Stochastic Partial Differential Equations

A Concise Course on Stochastic Partial Differential Equations A Concise Course on Stochastic Partial Differential Equations Michael Röckner Reference: C. Prevot, M. Röckner: Springer LN in Math. 1905, Berlin (2007) And see the references therein for the original

More information

In terms of measures: Exercise 1. Existence of a Gaussian process: Theorem 2. Remark 3.

In terms of measures: Exercise 1. Existence of a Gaussian process: Theorem 2. Remark 3. 1. GAUSSIAN PROCESSES A Gaussian process on a set T is a collection of random variables X =(X t ) t T on a common probability space such that for any n 1 and any t 1,...,t n T, the vector (X(t 1 ),...,X(t

More information

GAUSSIAN PROCESSES; KOLMOGOROV-CHENTSOV THEOREM

GAUSSIAN PROCESSES; KOLMOGOROV-CHENTSOV THEOREM GAUSSIAN PROCESSES; KOLMOGOROV-CHENTSOV THEOREM STEVEN P. LALLEY 1. GAUSSIAN PROCESSES: DEFINITIONS AND EXAMPLES Definition 1.1. A standard (one-dimensional) Wiener process (also called Brownian motion)

More information

9 Brownian Motion: Construction

9 Brownian Motion: Construction 9 Brownian Motion: Construction 9.1 Definition and Heuristics The central limit theorem states that the standard Gaussian distribution arises as the weak limit of the rescaled partial sums S n / p n of

More information

A Short Introduction to Diffusion Processes and Ito Calculus

A Short Introduction to Diffusion Processes and Ito Calculus A Short Introduction to Diffusion Processes and Ito Calculus Cédric Archambeau University College, London Center for Computational Statistics and Machine Learning c.archambeau@cs.ucl.ac.uk January 24,

More information

FE 5204 Stochastic Differential Equations

FE 5204 Stochastic Differential Equations Instructor: Jim Zhu e-mail:zhu@wmich.edu http://homepages.wmich.edu/ zhu/ January 20, 2009 Preliminaries for dealing with continuous random processes. Brownian motions. Our main reference for this lecture

More information

The concentration of a drug in blood. Exponential decay. Different realizations. Exponential decay with noise. dc(t) dt.

The concentration of a drug in blood. Exponential decay. Different realizations. Exponential decay with noise. dc(t) dt. The concentration of a drug in blood Exponential decay C12 concentration 2 4 6 8 1 C12 concentration 2 4 6 8 1 dc(t) dt = µc(t) C(t) = C()e µt 2 4 6 8 1 12 time in minutes 2 4 6 8 1 12 time in minutes

More information

Numerical methods for solving stochastic differential equations

Numerical methods for solving stochastic differential equations Mathematical Communications 4(1999), 251-256 251 Numerical methods for solving stochastic differential equations Rózsa Horváth Bokor Abstract. This paper provides an introduction to stochastic calculus

More information

Universal examples. Chapter The Bernoulli process

Universal examples. Chapter The Bernoulli process Chapter 1 Universal examples 1.1 The Bernoulli process First description: Bernoulli random variables Y i for i = 1, 2, 3,... independent with P [Y i = 1] = p and P [Y i = ] = 1 p. Second description: Binomial

More information

Brownian Motion and Conditional Probability

Brownian Motion and Conditional Probability Math 561: Theory of Probability (Spring 2018) Week 10 Brownian Motion and Conditional Probability 10.1 Standard Brownian Motion (SBM) Brownian motion is a stochastic process with both practical and theoretical

More information

Stochastic Processes. Winter Term Paolo Di Tella Technische Universität Dresden Institut für Stochastik

Stochastic Processes. Winter Term Paolo Di Tella Technische Universität Dresden Institut für Stochastik Stochastic Processes Winter Term 2016-2017 Paolo Di Tella Technische Universität Dresden Institut für Stochastik Contents 1 Preliminaries 5 1.1 Uniform integrability.............................. 5 1.2

More information

Bernardo D Auria Stochastic Processes /12. Notes. March 29 th, 2012

Bernardo D Auria Stochastic Processes /12. Notes. March 29 th, 2012 1 Stochastic Calculus Notes March 9 th, 1 In 19, Bachelier proposed for the Paris stock exchange a model for the fluctuations affecting the price X(t) of an asset that was given by the Brownian motion.

More information

Stochastic differential equation models in biology Susanne Ditlevsen

Stochastic differential equation models in biology Susanne Ditlevsen Stochastic differential equation models in biology Susanne Ditlevsen Introduction This chapter is concerned with continuous time processes, which are often modeled as a system of ordinary differential

More information

Discretization of SDEs: Euler Methods and Beyond

Discretization of SDEs: Euler Methods and Beyond Discretization of SDEs: Euler Methods and Beyond 09-26-2006 / PRisMa 2006 Workshop Outline Introduction 1 Introduction Motivation Stochastic Differential Equations 2 The Time Discretization of SDEs Monte-Carlo

More information

Exercises. T 2T. e ita φ(t)dt.

Exercises. T 2T. e ita φ(t)dt. Exercises. Set #. Construct an example of a sequence of probability measures P n on R which converge weakly to a probability measure P but so that the first moments m,n = xdp n do not converge to m = xdp.

More information

µ X (A) = P ( X 1 (A) )

µ X (A) = P ( X 1 (A) ) 1 STOCHASTIC PROCESSES This appendix provides a very basic introduction to the language of probability theory and stochastic processes. We assume the reader is familiar with the general measure and integration

More information

Lecture 17 Brownian motion as a Markov process

Lecture 17 Brownian motion as a Markov process Lecture 17: Brownian motion as a Markov process 1 of 14 Course: Theory of Probability II Term: Spring 2015 Instructor: Gordan Zitkovic Lecture 17 Brownian motion as a Markov process Brownian motion is

More information

Some Tools From Stochastic Analysis

Some Tools From Stochastic Analysis W H I T E Some Tools From Stochastic Analysis J. Potthoff Lehrstuhl für Mathematik V Universität Mannheim email: potthoff@math.uni-mannheim.de url: http://ls5.math.uni-mannheim.de To close the file, click

More information

1.1 Definition of BM and its finite-dimensional distributions

1.1 Definition of BM and its finite-dimensional distributions 1 Brownian motion Brownian motion as a physical phenomenon was discovered by botanist Robert Brown as he observed a chaotic motion of particles suspended in water. The rigorous mathematical model of BM

More information

JUSTIN HARTMANN. F n Σ.

JUSTIN HARTMANN. F n Σ. BROWNIAN MOTION JUSTIN HARTMANN Abstract. This paper begins to explore a rigorous introduction to probability theory using ideas from algebra, measure theory, and other areas. We start with a basic explanation

More information

Verona Course April Lecture 1. Review of probability

Verona Course April Lecture 1. Review of probability Verona Course April 215. Lecture 1. Review of probability Viorel Barbu Al.I. Cuza University of Iaşi and the Romanian Academy A probability space is a triple (Ω, F, P) where Ω is an abstract set, F is

More information

Filtrations, Markov Processes and Martingales. Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition

Filtrations, Markov Processes and Martingales. Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition Filtrations, Markov Processes and Martingales Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition David pplebaum Probability and Statistics Department,

More information

Stochastic Processes (Stochastik II)

Stochastic Processes (Stochastik II) Stochastic Processes (Stochastik II) Lecture Notes Zakhar Kabluchko University of Ulm Institute of Stochastics L A TEX-version: Judith Schmidt Vorwort Dies ist ein unvollständiges Skript zur Vorlesung

More information

Random Process Lecture 1. Fundamentals of Probability

Random Process Lecture 1. Fundamentals of Probability Random Process Lecture 1. Fundamentals of Probability Husheng Li Min Kao Department of Electrical Engineering and Computer Science University of Tennessee, Knoxville Spring, 2016 1/43 Outline 2/43 1 Syllabus

More information

Stochastic Differential Equations

Stochastic Differential Equations Chapter 5 Stochastic Differential Equations We would like to introduce stochastic ODE s without going first through the machinery of stochastic integrals. 5.1 Itô Integrals and Itô Differential Equations

More information

Poisson random measure: motivation

Poisson random measure: motivation : motivation The Lévy measure provides the expected number of jumps by time unit, i.e. in a time interval of the form: [t, t + 1], and of a certain size Example: ν([1, )) is the expected number of jumps

More information

Functional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals

Functional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals Functional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals Noèlia Viles Cuadros BCAM- Basque Center of Applied Mathematics with Prof. Enrico

More information

Stochastic process for macro

Stochastic process for macro Stochastic process for macro Tianxiao Zheng SAIF 1. Stochastic process The state of a system {X t } evolves probabilistically in time. The joint probability distribution is given by Pr(X t1, t 1 ; X t2,

More information

1. Stochastic Processes and filtrations

1. Stochastic Processes and filtrations 1. Stochastic Processes and 1. Stoch. pr., A stochastic process (X t ) t T is a collection of random variables on (Ω, F) with values in a measurable space (S, S), i.e., for all t, In our case X t : Ω S

More information

Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( )

Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( ) Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio (2014-2015) Etienne Tanré - Olivier Faugeras INRIA - Team Tosca November 26th, 2014 E. Tanré (INRIA - Team Tosca) Mathematical

More information

Lecture 12. F o s, (1.1) F t := s>t

Lecture 12. F o s, (1.1) F t := s>t Lecture 12 1 Brownian motion: the Markov property Let C := C(0, ), R) be the space of continuous functions mapping from 0, ) to R, in which a Brownian motion (B t ) t 0 almost surely takes its value. Let

More information

Some Terminology and Concepts that We will Use, But Not Emphasize (Section 6.2)

Some Terminology and Concepts that We will Use, But Not Emphasize (Section 6.2) Some Terminology and Concepts that We will Use, But Not Emphasize (Section 6.2) Statistical analysis is based on probability theory. The fundamental object in probability theory is a probability space,

More information

Brownian Motion. An Undergraduate Introduction to Financial Mathematics. J. Robert Buchanan. J. Robert Buchanan Brownian Motion

Brownian Motion. An Undergraduate Introduction to Financial Mathematics. J. Robert Buchanan. J. Robert Buchanan Brownian Motion Brownian Motion An Undergraduate Introduction to Financial Mathematics J. Robert Buchanan 2010 Background We have already seen that the limiting behavior of a discrete random walk yields a derivation of

More information

Lecture 19 : Brownian motion: Path properties I

Lecture 19 : Brownian motion: Path properties I Lecture 19 : Brownian motion: Path properties I MATH275B - Winter 2012 Lecturer: Sebastien Roch References: [Dur10, Section 8.1], [Lig10, Section 1.5, 1.6], [MP10, Section 1.1, 1.2]. 1 Invariance We begin

More information

The Wiener Itô Chaos Expansion

The Wiener Itô Chaos Expansion 1 The Wiener Itô Chaos Expansion The celebrated Wiener Itô chaos expansion is fundamental in stochastic analysis. In particular, it plays a crucial role in the Malliavin calculus as it is presented in

More information

p 1 ( Y p dp) 1/p ( X p dp) 1 1 p

p 1 ( Y p dp) 1/p ( X p dp) 1 1 p Doob s inequality Let X(t) be a right continuous submartingale with respect to F(t), t 1 P(sup s t X(s) λ) 1 λ {sup s t X(s) λ} X + (t)dp 2 For 1 < p

More information

Stochastic Integration and Stochastic Differential Equations: a gentle introduction

Stochastic Integration and Stochastic Differential Equations: a gentle introduction Stochastic Integration and Stochastic Differential Equations: a gentle introduction Oleg Makhnin New Mexico Tech Dept. of Mathematics October 26, 27 Intro: why Stochastic? Brownian Motion/ Wiener process

More information

1 Independent increments

1 Independent increments Tel Aviv University, 2008 Brownian motion 1 1 Independent increments 1a Three convolution semigroups........... 1 1b Independent increments.............. 2 1c Continuous time................... 3 1d Bad

More information

On pathwise stochastic integration

On pathwise stochastic integration On pathwise stochastic integration Rafa l Marcin Lochowski Afican Institute for Mathematical Sciences, Warsaw School of Economics UWC seminar Rafa l Marcin Lochowski (AIMS, WSE) On pathwise stochastic

More information

M4A42 APPLIED STOCHASTIC PROCESSES

M4A42 APPLIED STOCHASTIC PROCESSES M4A42 APPLIED STOCHASTIC PROCESSES G.A. Pavliotis Department of Mathematics Imperial College London, UK LECTURE 1 12/10/2009 Lectures: Mondays 09:00-11:00, Huxley 139, Tuesdays 09:00-10:00, Huxley 144.

More information

Selected Exercises on Expectations and Some Probability Inequalities

Selected Exercises on Expectations and Some Probability Inequalities Selected Exercises on Expectations and Some Probability Inequalities # If E(X 2 ) = and E X a > 0, then P( X λa) ( λ) 2 a 2 for 0 < λ

More information

Local vs. Nonlocal Diffusions A Tale of Two Laplacians

Local vs. Nonlocal Diffusions A Tale of Two Laplacians Local vs. Nonlocal Diffusions A Tale of Two Laplacians Jinqiao Duan Dept of Applied Mathematics Illinois Institute of Technology Chicago duan@iit.edu Outline 1 Einstein & Wiener: The Local diffusion 2

More information

Introduction to Diffusion Processes.

Introduction to Diffusion Processes. Introduction to Diffusion Processes. Arka P. Ghosh Department of Statistics Iowa State University Ames, IA 511-121 apghosh@iastate.edu (515) 294-7851. February 1, 21 Abstract In this section we describe

More information

STAT 331. Martingale Central Limit Theorem and Related Results

STAT 331. Martingale Central Limit Theorem and Related Results STAT 331 Martingale Central Limit Theorem and Related Results In this unit we discuss a version of the martingale central limit theorem, which states that under certain conditions, a sum of orthogonal

More information

{σ x >t}p x. (σ x >t)=e at.

{σ x >t}p x. (σ x >t)=e at. 3.11. EXERCISES 121 3.11 Exercises Exercise 3.1 Consider the Ornstein Uhlenbeck process in example 3.1.7(B). Show that the defined process is a Markov process which converges in distribution to an N(0,σ

More information

An Introduction to Entropy and Subshifts of. Finite Type

An Introduction to Entropy and Subshifts of. Finite Type An Introduction to Entropy and Subshifts of Finite Type Abby Pekoske Department of Mathematics Oregon State University pekoskea@math.oregonstate.edu August 4, 2015 Abstract This work gives an overview

More information

Bernardo D Auria Stochastic Processes /10. Notes. Abril 13 th, 2010

Bernardo D Auria Stochastic Processes /10. Notes. Abril 13 th, 2010 1 Stochastic Calculus Notes Abril 13 th, 1 As we have seen in previous lessons, the stochastic integral with respect to the Brownian motion shows a behavior different from the classical Riemann-Stieltjes

More information

Feller Processes and Semigroups

Feller Processes and Semigroups Stat25B: Probability Theory (Spring 23) Lecture: 27 Feller Processes and Semigroups Lecturer: Rui Dong Scribe: Rui Dong ruidong@stat.berkeley.edu For convenience, we can have a look at the list of materials

More information

Do stochastic processes exist?

Do stochastic processes exist? Project 1 Do stochastic processes exist? If you wish to take this course for credit, you should keep a notebook that contains detailed proofs of the results sketched in my handouts. You may consult any

More information

A NOTE ON STOCHASTIC INTEGRALS AS L 2 -CURVES

A NOTE ON STOCHASTIC INTEGRALS AS L 2 -CURVES A NOTE ON STOCHASTIC INTEGRALS AS L 2 -CURVES STEFAN TAPPE Abstract. In a work of van Gaans (25a) stochastic integrals are regarded as L 2 -curves. In Filipović and Tappe (28) we have shown the connection

More information

6. Brownian Motion. Q(A) = P [ ω : x(, ω) A )

6. Brownian Motion. Q(A) = P [ ω : x(, ω) A ) 6. Brownian Motion. stochastic process can be thought of in one of many equivalent ways. We can begin with an underlying probability space (Ω, Σ, P) and a real valued stochastic process can be defined

More information

White noise generalization of the Clark-Ocone formula under change of measure

White noise generalization of the Clark-Ocone formula under change of measure White noise generalization of the Clark-Ocone formula under change of measure Yeliz Yolcu Okur Supervisor: Prof. Bernt Øksendal Co-Advisor: Ass. Prof. Giulia Di Nunno Centre of Mathematics for Applications

More information

M5A42 APPLIED STOCHASTIC PROCESSES

M5A42 APPLIED STOCHASTIC PROCESSES M5A42 APPLIED STOCHASTIC PROCESSES Professor G.A. Pavliotis Department of Mathematics Imperial College London, UK LECTURE 1 06/10/2016 Lectures: Thursdays 14:00-15:00, Huxley 140, Fridays 10:00-12:00,

More information

HJB equations. Seminar in Stochastic Modelling in Economics and Finance January 10, 2011

HJB equations. Seminar in Stochastic Modelling in Economics and Finance January 10, 2011 Department of Probability and Mathematical Statistics Faculty of Mathematics and Physics, Charles University in Prague petrasek@karlin.mff.cuni.cz Seminar in Stochastic Modelling in Economics and Finance

More information

arxiv: v2 [math.pr] 27 Oct 2015

arxiv: v2 [math.pr] 27 Oct 2015 A brief note on the Karhunen-Loève expansion Alen Alexanderian arxiv:1509.07526v2 [math.pr] 27 Oct 2015 October 28, 2015 Abstract We provide a detailed derivation of the Karhunen Loève expansion of a stochastic

More information

Mean-square Stability Analysis of an Extended Euler-Maruyama Method for a System of Stochastic Differential Equations

Mean-square Stability Analysis of an Extended Euler-Maruyama Method for a System of Stochastic Differential Equations Mean-square Stability Analysis of an Extended Euler-Maruyama Method for a System of Stochastic Differential Equations Ram Sharan Adhikari Assistant Professor Of Mathematics Rogers State University Mathematical

More information

Lecture 4: Numerical Solution of SDEs, Itô Taylor Series, Gaussian Process Approximations

Lecture 4: Numerical Solution of SDEs, Itô Taylor Series, Gaussian Process Approximations Lecture 4: Numerical Solution of SDEs, Itô Taylor Series, Gaussian Process Approximations Simo Särkkä Aalto University Tampere University of Technology Lappeenranta University of Technology Finland November

More information

Stochastic Calculus (Lecture #3)

Stochastic Calculus (Lecture #3) Stochastic Calculus (Lecture #3) Siegfried Hörmann Université libre de Bruxelles (ULB) Spring 2014 Outline of the course 1. Stochastic processes in continuous time. 2. Brownian motion. 3. Itô integral:

More information

FE610 Stochastic Calculus for Financial Engineers. Stevens Institute of Technology

FE610 Stochastic Calculus for Financial Engineers. Stevens Institute of Technology FE610 Stochastic Calculus for Financial Engineers Lecture 3. Calculaus in Deterministic and Stochastic Environments Steve Yang Stevens Institute of Technology 01/31/2012 Outline 1 Modeling Random Behavior

More information

A D VA N C E D P R O B A B I L - I T Y

A D VA N C E D P R O B A B I L - I T Y A N D R E W T U L L O C H A D VA N C E D P R O B A B I L - I T Y T R I N I T Y C O L L E G E T H E U N I V E R S I T Y O F C A M B R I D G E Contents 1 Conditional Expectation 5 1.1 Discrete Case 6 1.2

More information

Lecture 12: Detailed balance and Eigenfunction methods

Lecture 12: Detailed balance and Eigenfunction methods Miranda Holmes-Cerfon Applied Stochastic Analysis, Spring 2015 Lecture 12: Detailed balance and Eigenfunction methods Readings Recommended: Pavliotis [2014] 4.5-4.7 (eigenfunction methods and reversibility),

More information

1. Stochastic Process

1. Stochastic Process HETERGENEITY IN QUANTITATIVE MACROECONOMICS @ TSE OCTOBER 17, 216 STOCHASTIC CALCULUS BASICS SANG YOON (TIM) LEE Very simple notes (need to add references). It is NOT meant to be a substitute for a real

More information

Stochastic Differential Equations.

Stochastic Differential Equations. Chapter 3 Stochastic Differential Equations. 3.1 Existence and Uniqueness. One of the ways of constructing a Diffusion process is to solve the stochastic differential equation dx(t) = σ(t, x(t)) dβ(t)

More information

Example 4.1 Let X be a random variable and f(t) a given function of time. Then. Y (t) = f(t)x. Y (t) = X sin(ωt + δ)

Example 4.1 Let X be a random variable and f(t) a given function of time. Then. Y (t) = f(t)x. Y (t) = X sin(ωt + δ) Chapter 4 Stochastic Processes 4. Definition In the previous chapter we studied random variables as functions on a sample space X(ω), ω Ω, without regard to how these might depend on parameters. We now

More information

Gaussian, Markov and stationary processes

Gaussian, Markov and stationary processes Gaussian, Markov and stationary processes Gonzalo Mateos Dept. of ECE and Goergen Institute for Data Science University of Rochester gmateosb@ece.rochester.edu http://www.ece.rochester.edu/~gmateosb/ November

More information

Homework # , Spring Due 14 May Convergence of the empirical CDF, uniform samples

Homework # , Spring Due 14 May Convergence of the empirical CDF, uniform samples Homework #3 36-754, Spring 27 Due 14 May 27 1 Convergence of the empirical CDF, uniform samples In this problem and the next, X i are IID samples on the real line, with cumulative distribution function

More information

Modeling with Itô Stochastic Differential Equations

Modeling with Itô Stochastic Differential Equations Modeling with Itô Stochastic Differential Equations 2.4-2.6 E. Allen presentation by T. Perälä 27.0.2009 Postgraduate seminar on applied mathematics 2009 Outline Hilbert Space of Stochastic Processes (

More information

Controlled Diffusions and Hamilton-Jacobi Bellman Equations

Controlled Diffusions and Hamilton-Jacobi Bellman Equations Controlled Diffusions and Hamilton-Jacobi Bellman Equations Emo Todorov Applied Mathematics and Computer Science & Engineering University of Washington Winter 2014 Emo Todorov (UW) AMATH/CSE 579, Winter

More information

Inference for Stochastic Processes

Inference for Stochastic Processes Inference for Stochastic Processes Robert L. Wolpert Revised: June 19, 005 Introduction A stochastic process is a family {X t } of real-valued random variables, all defined on the same probability space

More information

Tools of stochastic calculus

Tools of stochastic calculus slides for the course Interest rate theory, University of Ljubljana, 212-13/I, part III József Gáll University of Debrecen Nov. 212 Jan. 213, Ljubljana Itô integral, summary of main facts Notations, basic

More information

1 Brownian Local Time

1 Brownian Local Time 1 Brownian Local Time We first begin by defining the space and variables for Brownian local time. Let W t be a standard 1-D Wiener process. We know that for the set, {t : W t = } P (µ{t : W t = } = ) =

More information

SUMMARY OF RESULTS ON PATH SPACES AND CONVERGENCE IN DISTRIBUTION FOR STOCHASTIC PROCESSES

SUMMARY OF RESULTS ON PATH SPACES AND CONVERGENCE IN DISTRIBUTION FOR STOCHASTIC PROCESSES SUMMARY OF RESULTS ON PATH SPACES AND CONVERGENCE IN DISTRIBUTION FOR STOCHASTIC PROCESSES RUTH J. WILLIAMS October 2, 2017 Department of Mathematics, University of California, San Diego, 9500 Gilman Drive,

More information

Stochastic Processes

Stochastic Processes Introduction and Techniques Lecture 4 in Financial Mathematics UiO-STK4510 Autumn 2015 Teacher: S. Ortiz-Latorre Stochastic Processes 1 Stochastic Processes De nition 1 Let (E; E) be a measurable space

More information

Itô s formula. Samy Tindel. Purdue University. Probability Theory 2 - MA 539

Itô s formula. Samy Tindel. Purdue University. Probability Theory 2 - MA 539 Itô s formula Samy Tindel Purdue University Probability Theory 2 - MA 539 Mostly taken from Brownian Motion and Stochastic Calculus by I. Karatzas and S. Shreve Samy T. Itô s formula Probability Theory

More information

The Cameron-Martin-Girsanov (CMG) Theorem

The Cameron-Martin-Girsanov (CMG) Theorem The Cameron-Martin-Girsanov (CMG) Theorem There are many versions of the CMG Theorem. In some sense, there are many CMG Theorems. The first version appeared in ] in 944. Here we present a standard version,

More information

1 Math 241A-B Homework Problem List for F2015 and W2016

1 Math 241A-B Homework Problem List for F2015 and W2016 1 Math 241A-B Homework Problem List for F2015 W2016 1.1 Homework 1. Due Wednesday, October 7, 2015 Notation 1.1 Let U be any set, g be a positive function on U, Y be a normed space. For any f : U Y let

More information

Wiener Measure and Brownian Motion

Wiener Measure and Brownian Motion Chapter 16 Wiener Measure and Brownian Motion Diffusion of particles is a product of their apparently random motion. The density u(t, x) of diffusing particles satisfies the diffusion equation (16.1) u

More information

MATH 6605: SUMMARY LECTURE NOTES

MATH 6605: SUMMARY LECTURE NOTES MATH 6605: SUMMARY LECTURE NOTES These notes summarize the lectures on weak convergence of stochastic processes. If you see any typos, please let me know. 1. Construction of Stochastic rocesses A stochastic

More information

Lecture 21 Representations of Martingales

Lecture 21 Representations of Martingales Lecture 21: Representations of Martingales 1 of 11 Course: Theory of Probability II Term: Spring 215 Instructor: Gordan Zitkovic Lecture 21 Representations of Martingales Right-continuous inverses Let

More information

Nonparametric Bayesian Methods - Lecture I

Nonparametric Bayesian Methods - Lecture I Nonparametric Bayesian Methods - Lecture I Harry van Zanten Korteweg-de Vries Institute for Mathematics CRiSM Masterclass, April 4-6, 2016 Overview of the lectures I Intro to nonparametric Bayesian statistics

More information

SMSTC (2007/08) Probability.

SMSTC (2007/08) Probability. SMSTC (27/8) Probability www.smstc.ac.uk Contents 12 Markov chains in continuous time 12 1 12.1 Markov property and the Kolmogorov equations.................... 12 2 12.1.1 Finite state space.................................

More information

(B(t i+1 ) B(t i )) 2

(B(t i+1 ) B(t i )) 2 ltcc5.tex Week 5 29 October 213 Ch. V. ITÔ (STOCHASTIC) CALCULUS. WEAK CONVERGENCE. 1. Quadratic Variation. A partition π n of [, t] is a finite set of points t ni such that = t n < t n1

More information

Stochastic Processes II/ Wahrscheinlichkeitstheorie III. Lecture Notes

Stochastic Processes II/ Wahrscheinlichkeitstheorie III. Lecture Notes BMS Basic Course Stochastic Processes II/ Wahrscheinlichkeitstheorie III Michael Scheutzow Lecture Notes Technische Universität Berlin Sommersemester 218 preliminary version October 12th 218 Contents

More information

Stochastic Calculus February 11, / 33

Stochastic Calculus February 11, / 33 Martingale Transform M n martingale with respect to F n, n =, 1, 2,... σ n F n (σ M) n = n 1 i= σ i(m i+1 M i ) is a Martingale E[(σ M) n F n 1 ] n 1 = E[ σ i (M i+1 M i ) F n 1 ] i= n 2 = σ i (M i+1 M

More information

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed

More information

Exercises in stochastic analysis

Exercises in stochastic analysis Exercises in stochastic analysis Franco Flandoli, Mario Maurelli, Dario Trevisan The exercises with a P are those which have been done totally or partially) in the previous lectures; the exercises with

More information

LECTURE 2: LOCAL TIME FOR BROWNIAN MOTION

LECTURE 2: LOCAL TIME FOR BROWNIAN MOTION LECTURE 2: LOCAL TIME FOR BROWNIAN MOTION We will define local time for one-dimensional Brownian motion, and deduce some of its properties. We will then use the generalized Ray-Knight theorem proved in

More information

Lecture 1: Brief Review on Stochastic Processes

Lecture 1: Brief Review on Stochastic Processes Lecture 1: Brief Review on Stochastic Processes A stochastic process is a collection of random variables {X t (s) : t T, s S}, where T is some index set and S is the common sample space of the random variables.

More information

Statistics 992 Continuous-time Markov Chains Spring 2004

Statistics 992 Continuous-time Markov Chains Spring 2004 Summary Continuous-time finite-state-space Markov chains are stochastic processes that are widely used to model the process of nucleotide substitution. This chapter aims to present much of the mathematics

More information

The Codimension of the Zeros of a Stable Process in Random Scenery

The Codimension of the Zeros of a Stable Process in Random Scenery The Codimension of the Zeros of a Stable Process in Random Scenery Davar Khoshnevisan The University of Utah, Department of Mathematics Salt Lake City, UT 84105 0090, U.S.A. davar@math.utah.edu http://www.math.utah.edu/~davar

More information

E. Vitali Lecture 6. Brownian motion

E. Vitali Lecture 6. Brownian motion Markov chains, presented in the previous chapter, belong to a wide class of mathematical objects, the stochastic processes, which will introduced in the present chapter. Stochastic processes form the basis

More information