Chapter 4: Monte-Carlo Methods

Size: px
Start display at page:

Download "Chapter 4: Monte-Carlo Methods"

Transcription

1 Chapter 4: Monte-Carlo Methods A Monte-Carlo method is a technique for the numerical realization of a stochastic process by means of normally distributed random variables. In financial mathematics, it is used for the computation of the expectation of option prices in situations where explicit solutions like the Black-Scholes formula are not accessible (e.g., exotic options). For simplicity, we consider a European plain-vanilla put option on a stock S t whose value behaves according to the geometric Brownian motion ( ) d S t = r S t dt + σ S t dw t with a risk-free interest rate r 0, constant volatility σ > 0, and a Wiener process W t. We recall that the option price V(S t,t) at t = 0 is given by the discounted expectation ( ) V(S 0,0) = exp( rt) E(V(S T,T)).

2 The Four Basic Steps of the Monte-Carlo Simulation The idea behind the Monte-Carlo method is based on the approximation of the expectation E(V(S T,T)) in ( ) by the simulation of M pathes {S t 0 < t < T} of the value of the stock: Step 1: Simulation of the pathes For M independent pathes, compute the solution S (k) t,1 k M, of the geometric Brownian motion ( ). This requires the simulation of M independent realizations of the Wiener process W t and the numerical solution of the stochastic differential equation for the respective pathes. Step 2: Computation of the payoff function For 1 k M, compute the payoff function associated with the path S (k) t : V (k) T = (K S (k) T ) +. Step 3: Computation of an estimate of the expectation According to the law of large numbers, the arithmetic mean of equally distributed, independent random variables converges almost surely to the expectation. Hence, an appropriate estimate of the expectation in ( ) is given by V (k) T = (K S (k) T ) +. Step 4: Computation of an estimate of the option price Compute an approximation of the option price by ˆV = exp( rt) Ê(V T ).

3 4.1 Generation of Equally Distributed Random Numbers For the simulation of a Wiener process, we use standardized normally distributed random numbers Z N(0,1) in order to compute the increments W = Z t. Definition 4.1 (Equally Distributed Random Variables) (i) A random variable X is called equally distributed on [a,b] lr, (notation: X U[a,b]), if its density function is given by f(x) = 1/(b a), x [a, b]. (ii) A sequence {X n } n ln of random variables is called F-distributed, if the X n are independent realizations of random numbers with the distribution function F. In the sequel, we will consider the following random number generators: Linear congruential generators The generator RANDU Fibonacci generators

4 4.1.1 Linear Congruential Generators Linear congruential generators are obtained by the following algorithm: Algorithm 4.1: Linear Congruential Generators Given M ln,a,b {0,1,...,M 1},a 0, and a seed X 0 {0,1,...,M 1}, compute For i = 1,2,... X i := (a X i 1 + b) mod M, U i := X i /M. where a mod M := M (a/m a/m ) with a/m : largest integer less or equal a/m. Lemma 4.1 (Properties of Linear Congruential Generators) The random numbers generated by Algorithm 4.1 have the properties (i) X i {0,1,...,M 1}. (ii) The X i are periodic with period less or equal M. (iii) If b = 0 (multiplicative congruential generator), we must exclude the seed X 0 = 0. (iv) If a = 1, we obtain X i = (X 0 + ib) mod M (too easy to predict!).

5 Analysis of Linear Congruential Generators The random numbers X i and U i,i 1, obtained by the linear congruential generator, can be written as m-tuples (X i,x i+1,...,x i+m 1 ) or points (U i,...,u i+m 1 ) [0,1) m. They are situated on (m 1)-dimensional hyperplanes which can be constructed as the following analysis in case m = 2 reveals: X i = (a X i 1 + b) mod M X i = a X i 1 + b k M, k M a X i 1 + b < (k + 1) M. For an arbitrary tuple (z 0,z 1 ) Z 2 we obtain: ( ) z 0 X i 1 + z 1 X i = z 0 X i 1 + z 1 (a X i 1 + b k M) = = X i 1 (z 0 + a z 1 ) + z 1 b z 1 k M = = M (x i 1 z 0 + a z 1 M z 1 k) }{{} =: c + z 1 b. Division by M yields the equation of a straight line in the (U i 1,U i )-plane: ( ) z 0 U i 1 + z 1 U i = c + z 1 b M 1.

6 Analysis of Linear Congruential Generators Question: Does there exist a tuple (z 0,z 1 ) Z 2 such that the intersection of the straight lines with the square [0,1) 2 is non-empty for only a few lines which would violate the requirement of uniform distributions of the points? If we choose (z 0,z 1 ) Z 2 and z 0 +a z 1 = 0 mod M, then c Z. Solving for c in ( ) results in the maximum interval c min c c max such that the associated straight line ( ) has non-zero intersection with [0,1) 2. Figure 4.1: M = 2048, a = 1229, b = 1, X 0 = 1. The figure displays the first 499 points (U i 1,U i ) obtained by the linear congruential generator. These points are lying on 5 parallel straight lines.

7 4.1.2 The Random Number Generator RANDU The generator RANDU is based on a multiplicative congruential method with the data M = 2 31, a = , b = 0. Figure 4.2 suggests that the points (U i 1,U i ) are randomly distributed in the plane. However, Figure 4.3 displays the tuples (U i 2,U i 1,U i ), and we observe that they are lying on only 15 different planes in the unit cube [0,1) 3 which is a serious disadvantage in many applications. Figure 4.2 Figure 4.3

8 4.1.3 Fibonacci Generators Fibonacci generators use the recursively given Fibonacci sequence according to where X 1,X 2 are given seeds. X i := (X i 1 + X i ) mod M, U i := X i /M, i 2, Figure 4.4 displays the first 2000 points (U i 1,U i ) obtained by the Fibonacci generator for M = 2179 and X 1 = X 2 = 1. Significantly less than 2000 points are seen in the unit square due to the fact that the sequence {U i } i ln repeats any 197 entries. Therefore, this Fibonacci generator is less suited for random number generation. Figure 4.4

9 Better and widely used random number generators are the so-called lagged Fibonacci generators: Algorithm 4.2: Lagged Fibonacci Generators Given M ln, µ, ν ln and initial values X 1,...,X max(µ,ν), for i max{µ, ν} compute X i := (X i µ + X i ν ) mod M, U i := X i /M. For M = 2048, µ = 5, ν = 17, Figure 4.5 shows the first 2000 points (U i 1,U i ) obtained by the lagged Fibonacci generator. Figure 4.5

10 4.2 Transformation of Random Variables Normally distributed random variables can be obtained by a transformation of equally distributed random variables using either the inversion of the distribution function or a transformation between random variables Inversion of the Distribution Function Theorem 4.1 (Inversion of the Distribution Function) Assume that U U[0,1] is an equally distributed random variable and assume further that F is a continuous, strongly monotone distribution function. Then, the random variable F 1 (U) is F-distributed. Proof. Since P(U ξ) = ξ, ξ [0,1], we have P(F 1 (U) x) = P(U F(x)) = F(x). Remark: For the normal distribution Φ there are no explicit expressions neither for Φ nor for Φ 1 which means that the nonlinear equation Φ(x) = u must be inverted numerically by, e.g., the application of a Newton-like method. However, for u 1 the problem is ill-conditioned.

11 4.2.2 Transformation of Random Variables Theorem 4.2 (Transformation of Random Variables) Assume that X is a random variable with density function f on A := {x lr n f(x) > 0}. Assume further that the mapping h : A B := h(a) is continuously invertible. Then, the random variable Y = h(x) has the density function where J h 1(y) is the Jacobian of h 1 in y. g(y) := f(h 1 (y)) det J h 1(y), y B, Proof. Applying the transformation rules for multiple integrals, we obtain P(Y = h(x) B) = P(X h 1 (B)) = h 1 (B) f(x) dx = f(h 1 (y)) det J h 1(y) dy. B

12 4.2 Normally Distributed Random Variables The Box-Muller Algorithm We apply Theorem 4.2 in case n = 2, A = S := [0,1] 2, f 1 on S, and h : S lr 2 given by y 1 = h 1 (x 1,x 2 ) = 2 ln(x 1 ) cos(2πx 2 ), y 2 = h 2 (x 1,x 2 ) = 2 ln(x 1 ) sin(2πx 2 ). Since y y 2 2 = 2 ln(x 1 ) and y 2 /y 1 = tan(2πx 2 ), the inverse function can be easily computed x 1 = exp( 1 2 (y2 1 + y 2 2)), x 2 = 1 2π arctan(y 2 y 1 ) with the determinant det J h 1(y) of the Jacobian J h 1(y) given by det J h 1(y) = 1 2π exp( 1 2 (y2 1 + y 2 2)) which is the density of the standard normal distribution in lr 2.

13 Algorithm 4.3: Box-Muller Algorithm Step 1: Generate U 1 U[0,1] and U 2 U[0,1]. Step 2: Set Θ := 2π U 2 and ρ := ln(u 1 ). Step 3: Z 1 := ρ cos(θ) and Z 2 := ρ sin(θ) are normally distributed random variables.. The histogram of Z 1 in Figure 4.6 based on random numbers shows that the algorithm indeed approximately produces normally distributed random variables. Figure 4.6

14 Numerische Mathematik für WiM Marsaglia s Polar Method Marsaglia s variant of the Box-Muller algorithm avoids the evalu For U 1,U 2 U[0,1], we obviously have W i := 2U i 1 [ 1, +1],1 a point in the plane. Only points inside the unit disk K := {(W 1,W 2 ) W W 2 2 < 1} are accepted. A Transformation K S := [0,1] 2 is provided by m x 1 x 2 = Algorithm 4.4: Marsaglia s Polar Method W1 2 + W π arctan(w 2 W 1 ) Step 1: Generate U 1,U 2 U[0,1] and W i := 2U i 1, 1 i 2, a Step 2: Z 1 := W 1 2ln(W W 2 2)//(W W 2 2) and Z 2 := W 2 2 normally distributed random variables. Remark: Marsaglia s polar method is more efficient than the Bo.

15 4.4 Correlated Random Variables We consider a random variable Z = (Z 1,...,Z n ) with density function f. We further denote by µ = E(Z) = (E(Z 1 ),...,E(Z n )) the expectation and by Σ = (Σ ij ) n i,j=1 with Σ ij = (Cov Z) ij := E((Z i µ i )(Z j µ j )) the associated covariance matrix which is symmetric, positive definite. We want to construct an N(µ,Σ)-distributed random variable Y. Theorem 4.3 (Correlated Random Variables) Under the above notations let Σ = LL T be the Cholesky decomposition of Σ with a lower triangular matrix L lr n n. Then, Y = µ + LZ is an N(µ,Σ)-distributed random variable with density function 1 1 f(x) = (2π) n/2 (det Σ) exp( 1 1/2 2 (x µ)t Σ 1 (x µ)).

16 Proof. Setting x = Lz and observing dx = det L dz, we obtain f(z) dz = 1 z (2π) n/2 exp( zt 2 ) dz = 1 x) T L 1 x (2π) n/2 exp( (L 1 ) dz = 2 = 1 (LL T ) 1 x (2π) n/2 exp( xt ) dz = 2 1 Σ 1 x (2π) n/2 det L exp( xt ) dx = 2 = 1 Σ 1 x (2π) n/2 (det Σ) 1/2 exp( xt ) dx. 2 This shows that X = LZ is N(0,Σ)-distributed which readily gives the assertion.

17 Algorithm 4.5: Correlated Random Variables Step 1: Compute the Cholesky decomposition Σ = LL T. Step 2: Compute Z N(0,1) component-wise, i.e., Z i N(0,1), 1 i n, e.g., by Marsaglia s polar algorithm. Step 3: Y = µ + LZ is N(µ, Σ)-distributed. MATLAB program correlated.m randn( state,1) Sigma = [543;454;345] mu = [ 5010] ;N = 10000; L = chol(sigma); X = zeros(3,n); for i = 1 : N X(:.i) = mu + L randn(3,1); end

18 4.5 Sequences of Random Numbers with Low Discrepancy As a measure for the quality of uniform distribution of random numbers we introduce the notion of discrepancy and give examples of sequences of random numbers with low discrepancy. A related issue is that to improve convergence properties in relevant applications such as the numerical integration of high-dimensional integrals Monte Carlo Integration Given a bounded domain Ω lr m and an integrable function f : Ω lr, the Monte Carlo integration provides an approximation of the m-dimensional integral of f over Ω by the weighted sum of the values of the integrand f at randomly chosen points x i,1 i N, in the domain of integration Ω according to I := f(x) dx vol(ω) N f(x i ) =: Q N. Ω lr m N i=1

19 If the random variables x 1,...,x N Ω are independent and uniformly distributed, the law of large numbers implies Q N vol(ω) E(f) = Ω f(x) dx as N. Hence, for the variance of the error ε N := I Q N we obtain Var(ε N ) = E(ε 2 N) E(ε N ) 2 = σ2 (f) N vol(ω)2, σ 2 (f) = f(x) 2 dx ( f(x) dx) 2. Ω Ω The central limit theorem implies the following convergence result for the standard deviation Var(εN ) = O(N 1/2 ) as N, which means slow convergence (improvement of the accuracy by a factor of 10 requires to increase the computational cost by a factor of 100). Another disadvantage is the lack of a strict error bound. Before we discuss methods of variance reduction in section 4.6, we will address the issue to choose the random points x i,1 i N, in such a way that with increasing N the accuracy gets better but clustering is avoided.

20 4.5.2 Discrepancy The idea behind discrepancy is that for a uniformly distributed point set {x 1,...,x N } [0,1] m the fraction of the points lying within some quadrilateral Q [0,1] m should correspond to the volume of the quadrilateral, i.e., card({x i,1 i N x i Q}) N Definition 4.2 (Discrepancy and Star Discrepancy) The discrepancy of a point set {x 1,...,x N } [0,1] m is defined by D N := sup card({x i,1 i N x i Q}) Q [0,1] m N vol(q) vol([0,1] m ). vol(q), where the supremum is taken with respect to all quadrilaterals Q [0,1] m. The star discrepancy D N is defined in the same way, but with the supremum taken with respect to those quadrilaterals Q [0,1] m for which one corner is in the origin, i.e., Q = m i=1 [0,y i ] where (y 1,...,y m ) denotes the diagonally opposite corner.

21 Properties of Discrepancy and Star Discrepancy Theorem 4.4 (Properties of Discrepancy and Star Discrepancy) Let D N and D N be the discrepancy and star discrepancy of a point set {x i } ln,x i lr m,1 i N. Then, there holds Proof. The proof is left as an exercise. Upper Bound for Monte Carlo Integration D N D N 2 m D N. Theorem 4.5 (Theorem of Hlawka and Koksma) Assume that Ω := [0,1] m,m ln, and f : [0,1] m lr is a function of bounded variation BV(f). Then, for the error ε N in the Monte Carlo integration there holds ε N BV(f) D N.

22 Sequences of Low Discrepancy Definition 4.3 (Sequences of Low Discrepancy) A sequence {x i } N i=1 of points or numbers x i lr m,1 i N, is called a sequence of low discrepancy, if there exists a constant C m > 0 such that for all N ln: Remarks: D N C m (ln(n)) m (i) For not too large m ln we have that approximately D N O(N 1 ). Compare this with D N = O(N 1/2 ) for sequences of random numbers in general. (ii) Let U i {0,1/M,..., (M 1)/M} be random numbers obtained by a linear congruential generator. These are not even uniformly distributed, since with Q := [1/(2M + 2),1/(M + 1)] D N card({u i U i Q }) vol(q ) = vol(q 1 ) = N 2M + 2. N.

23 Van der Corput Sequence Remark: The sequence {X (N) i } N i=1 with x (N) i = i/n,1 i N, obviously satisfies D N = 1/N and thus is of low discrepancy. However, it is not suited for Monte Carlo integration, since for each N a new sequence has to be set up. For increasing N, it is more efficient to use already computed numbers. Examples of sequences of low discrepancy: (i) Van der Corput sequence Let i = (d j...d 0 ) 2 := j k=0 d k 2 k be the representation of the integer i ln in fixed point arithmetics with respect to the basis B = 2. Then, the sequence {x i } ln obtained by bit inversion according to is called van der Corput sequence. x i = (.d 0...d j ) 2 := j k=0 d k 2 (k+1)

24 Radical-Inverse Function and the Halton Sequence (ii) Radical-Inverse Function The radical-inverse function Φ B allows to define van der Corput sequences {x i } N i=1 with respect to an arbitrary basis B. Using the same notation as in (i), they are defined by (iii) Halton Sequence Φ B (i) := j k=0 d k B (k+1), x i := Φ B (i). Assume that p 1,...,p m are pairwise prime integers. The Halton sequence is defined as the sequence of vectors x i = {Φ p1 (i),...,φ pm (i)}, i ln.

25 Properties of van der Corput Sequences Theorem 4.6 (Properties of van der Corput Sequences) For the discrepancy of van der Corput sequences there holds lim sup N N D N ln(n) = B 2 4(B+1) ln(b) B 1 4 ln(b), B even,, B odd.

26 4.6 Numerical Integration of Stochastic Differential Equations The Euler-Maruyama Method In case of an ordinary differential equation x (t) = a(t,x(t)), T > 0, where a C(lR +, lr) satisfies a Lipschitz condition in the second argument, the explicit Euler method with uniform step-size h > 0 is given by y i+1 = y i + h a(t i,y i ), t i := i h, i 0. It is well known that it has the order of convergence 1, i.e. sup y i x(t i ) C h. i 0 For the stochastic differential equation dx t = a(t,x t ) dt + b(t,x t ) dw t the counterpart of the explicit Euler method is the Euler-Maruyama method y i+1 = y i + h a(t i,y i ) + b(t i,y i ) W i, W i := W ti+1 W ti.

27 4.6.2 Strong and weak convergence Definition 4.4 (Strong and Weak Convergence) Assume that x h t is an approximation of the solution x t of the stochastic differential equation dx t = a(t,x t ) dt + b(t,x t ) dw t. (i) The approximation x h T is said to converge strongly to x T of order γ > 0, if there exists a constant C > 0, such that for all sufficiently small h > 0 E( x T x h T ) C h γ. (ii) The approximation x h T is said to converge weakly to x T of order κ > 0, if there exists a constant C > 0, such that for all sufficiently small h > 0 E(x T ) E(x h T) C h κ. Theorem 4.7 (Strong and Weak Convergence of the Euler-Maruyama Method) The Euler-Maruyama method is strongly convergent of order γ = 0.5 and weakly convergent of order κ = 1. Proof. We refer to P. Kloeden and E. Platen; Numerical Solution of Stochastic Differential Equations. Springer, Berlin-Heidelberg-New York, 1999

28 4.6.3 The Method of Milstein We recall Itô s Lemma which states: If X t is an Itô process, then for sufficiently smooth f we have that f = f(t,x t ) is an Itô process as well which satisfies df = ( f t + a f x b2 2 f f x2) dt + b x dw t. The associated integral form is t ( ) f(x t ) = f(x t0 ) + (f (x s ) a(x s ) + 1 t 0 2 f (x s ) b(x s ) 2 ) ds + f (x s ) b(x s ) dw s. t 0 In particular, for f(x) = x we obtain ( ) x t = x t0 + a(x s ) ds + b(x s ) dw s. t 0 t 0 Choosing f = a and f = b in ( ) and inserting it into ( ) results in x t = x t0 + t t s t (a(x t0 ) + (a a + 1 t 0 t 0 2 a b 2 ) dz + a b dw z ) ds + t 0 + (b(x t0 ) + (b a + 1 t 0 t 0 2 b b 2 ) dz + b b dw z ) dw s, t 0 where a = a(x z ), b = b(x z ) etc. s s t s t

29 Modifying the double integral w.r.t. dw z dw s by replacing the integrand b (x z )b(x z ) b (x t0 )b(x t0 ), we can write x t = x t0 + a(x t0 ) (t t 0 ) + b(x t0 ) dw s + b (x t0 ) b(x t0 ) t 0 where the remainder is of the (higher) order R = O(h 3/2 ). The double integral can be evaluated as follows: t s t 0 t 0 dw z dw s = t t 0 (W s W t0 ) dw s = = 1 2 (W2 t Wt 2 0 ) t t 0 2 This leads to the method of Milstein t t t s W s dw s W t0 dw s = t 0 t 0 t t 0 t 0 dw z dw s + R, W t0 (W t W t0 ) = 1 2 ((W2 t W 2 t 0 ) t t 0 2 ). y i+1 = y i + h a(t i,y i ) + b(t i,y i ) W b (t i,y i ) b(t i,y i ) (( W) 2 h), where W := Z h, Z N(0,1). Theorem 4.8 (Convergence of the Method of Milstein) The method of Milstein is strongly convergent of order γ = 1.

30 4.7 Reduction of Variances The slow convergence of Monte-Carlo simulations can be explained as follows: We assume that θ n = n k=1 Φ(X k )/n is some stochastic approximation of a stochastic integral θ = E(Φ(x)), where Φ(x) := g(x)/f(x) and X k,1 k n, are independent samples of an F-distributed random variable with F = f. Under the assumption E(Φ(X k )) = θ, Var(Φ(X k )) = σ 2, 1 k n, it follows that n E(θ n ) = 1 E(Φ(X k )) = θ, Var(θ n ) = 1 n Var(Φ(X n k=1 n 2 k )) = σ2 k=1 n. Using the Chebychev inequality for square integrable random variables Y P( Y E(Y) ) δ) Var(Y) δ 2, δ := σ/ ε n, we obtain the error estimate P( θ n θ σ ) ε = P( θ n θ < σ ) 1 ε. εn εn This means that the error gets the smaller the larger the number n of samples is chosen. But in order to reduce the error by a factor of 10, the number of samples must be increased by a factor of 100.

31 4.7.1 Antithetic Variables Theorem 4.9 (Variance reduction by an Antithetic Variable) Assume that the approximation θ n has been generated by a random variable Z N(0,1) and generate an approximation θ n by means of Z N(0,1) such that Var(θ n ) = Var(θ n). Introduce the antithetic variable ˆθ n according to ˆθ n := 1 2 (θ n + θ n). Then, there holds Proof. The fundamental relation implies Taking into account that we thus obtain Var(ˆθ n ) < Var(θ n ). Var(X ± Y) = Var(X) + Var(Y) ± 2 Cov(X, Y) Var(ˆθ n ) = 1 4 Var(θ n + θ n) = 1 4 (Var(θ n) + Var(θ n) + 2 Cov(θ n, θ n)). Cov(X,Y) 1 2 (Var(X) + Var(Y)), Var(ˆθ n ) 1 4 (Var(θ n) + Var(θ n)), if Cov(θ n, θ n) 0, Var(ˆθ n ) < 1 4 (Var(θ n) + Var(θ n)), if Cov(θ n, θ n) > 0, which gives the assertion observing Var(θ n ) = Var(θ n).

A numerical method for solving uncertain differential equations

A numerical method for solving uncertain differential equations Journal of Intelligent & Fuzzy Systems 25 (213 825 832 DOI:1.3233/IFS-12688 IOS Press 825 A numerical method for solving uncertain differential equations Kai Yao a and Xiaowei Chen b, a Department of Mathematical

More information

1. Stochastic Processes and filtrations

1. Stochastic Processes and filtrations 1. Stochastic Processes and 1. Stoch. pr., A stochastic process (X t ) t T is a collection of random variables on (Ω, F) with values in a measurable space (S, S), i.e., for all t, In our case X t : Ω S

More information

Lecture 15 Random variables

Lecture 15 Random variables Lecture 15 Random variables Weinan E 1,2 and Tiejun Li 2 1 Department of Mathematics, Princeton University, weinan@princeton.edu 2 School of Mathematical Sciences, Peking University, tieli@pku.edu.cn No.1

More information

Discretization of SDEs: Euler Methods and Beyond

Discretization of SDEs: Euler Methods and Beyond Discretization of SDEs: Euler Methods and Beyond 09-26-2006 / PRisMa 2006 Workshop Outline Introduction 1 Introduction Motivation Stochastic Differential Equations 2 The Time Discretization of SDEs Monte-Carlo

More information

Some Terminology and Concepts that We will Use, But Not Emphasize (Section 6.2)

Some Terminology and Concepts that We will Use, But Not Emphasize (Section 6.2) Some Terminology and Concepts that We will Use, But Not Emphasize (Section 6.2) Statistical analysis is based on probability theory. The fundamental object in probability theory is a probability space,

More information

A Class of Fractional Stochastic Differential Equations

A Class of Fractional Stochastic Differential Equations Vietnam Journal of Mathematics 36:38) 71 79 Vietnam Journal of MATHEMATICS VAST 8 A Class of Fractional Stochastic Differential Equations Nguyen Tien Dung Department of Mathematics, Vietnam National University,

More information

5 December 2016 MAA136 Researcher presentation. Anatoliy Malyarenko. Topics for Bachelor and Master Theses. Anatoliy Malyarenko

5 December 2016 MAA136 Researcher presentation. Anatoliy Malyarenko. Topics for Bachelor and Master Theses. Anatoliy Malyarenko 5 December 216 MAA136 Researcher presentation 1 schemes The main problem of financial engineering: calculate E [f(x t (x))], where {X t (x): t T } is the solution of the system of X t (x) = x + Ṽ (X s

More information

Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( )

Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( ) Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio (2014-2015) Etienne Tanré - Olivier Faugeras INRIA - Team Tosca October 22nd, 2014 E. Tanré (INRIA - Team Tosca) Mathematical

More information

Variance Reduction Techniques for Monte Carlo Simulations with Stochastic Volatility Models

Variance Reduction Techniques for Monte Carlo Simulations with Stochastic Volatility Models Variance Reduction Techniques for Monte Carlo Simulations with Stochastic Volatility Models Jean-Pierre Fouque North Carolina State University SAMSI November 3, 5 1 References: Variance Reduction for Monte

More information

Stochastic differential equation models in biology Susanne Ditlevsen

Stochastic differential equation models in biology Susanne Ditlevsen Stochastic differential equation models in biology Susanne Ditlevsen Introduction This chapter is concerned with continuous time processes, which are often modeled as a system of ordinary differential

More information

QMC methods in quantitative finance. and perspectives

QMC methods in quantitative finance. and perspectives , tradition and perspectives Johannes Kepler University Linz (JKU) WU Research Seminar What is the content of the talk Valuation of financial derivatives in stochastic market models using (QMC-)simulation

More information

Monte-Carlo MMD-MA, Université Paris-Dauphine. Xiaolu Tan

Monte-Carlo MMD-MA, Université Paris-Dauphine. Xiaolu Tan Monte-Carlo MMD-MA, Université Paris-Dauphine Xiaolu Tan tan@ceremade.dauphine.fr Septembre 2015 Contents 1 Introduction 1 1.1 The principle.................................. 1 1.2 The error analysis

More information

Nested Uncertain Differential Equations and Its Application to Multi-factor Term Structure Model

Nested Uncertain Differential Equations and Its Application to Multi-factor Term Structure Model Nested Uncertain Differential Equations and Its Application to Multi-factor Term Structure Model Xiaowei Chen International Business School, Nankai University, Tianjin 371, China School of Finance, Nankai

More information

Brownian Motion. An Undergraduate Introduction to Financial Mathematics. J. Robert Buchanan. J. Robert Buchanan Brownian Motion

Brownian Motion. An Undergraduate Introduction to Financial Mathematics. J. Robert Buchanan. J. Robert Buchanan Brownian Motion Brownian Motion An Undergraduate Introduction to Financial Mathematics J. Robert Buchanan 2010 Background We have already seen that the limiting behavior of a discrete random walk yields a derivation of

More information

UNCERTAINTY FUNCTIONAL DIFFERENTIAL EQUATIONS FOR FINANCE

UNCERTAINTY FUNCTIONAL DIFFERENTIAL EQUATIONS FOR FINANCE Surveys in Mathematics and its Applications ISSN 1842-6298 (electronic), 1843-7265 (print) Volume 5 (2010), 275 284 UNCERTAINTY FUNCTIONAL DIFFERENTIAL EQUATIONS FOR FINANCE Iuliana Carmen Bărbăcioru Abstract.

More information

Formulas for probability theory and linear models SF2941

Formulas for probability theory and linear models SF2941 Formulas for probability theory and linear models SF2941 These pages + Appendix 2 of Gut) are permitted as assistance at the exam. 11 maj 2008 Selected formulae of probability Bivariate probability Transforms

More information

Numerical methods for solving stochastic differential equations

Numerical methods for solving stochastic differential equations Mathematical Communications 4(1999), 251-256 251 Numerical methods for solving stochastic differential equations Rózsa Horváth Bokor Abstract. This paper provides an introduction to stochastic calculus

More information

Gaussian, Markov and stationary processes

Gaussian, Markov and stationary processes Gaussian, Markov and stationary processes Gonzalo Mateos Dept. of ECE and Goergen Institute for Data Science University of Rochester gmateosb@ece.rochester.edu http://www.ece.rochester.edu/~gmateosb/ November

More information

Solution for Problem 7.1. We argue by contradiction. If the limit were not infinite, then since τ M (ω) is nondecreasing we would have

Solution for Problem 7.1. We argue by contradiction. If the limit were not infinite, then since τ M (ω) is nondecreasing we would have 362 Problem Hints and Solutions sup g n (ω, t) g(ω, t) sup g(ω, s) g(ω, t) µ n (ω). t T s,t: s t 1/n By the uniform continuity of t g(ω, t) on [, T], one has for each ω that µ n (ω) as n. Two applications

More information

The Multivariate Normal Distribution. In this case according to our theorem

The Multivariate Normal Distribution. In this case according to our theorem The Multivariate Normal Distribution Defn: Z R 1 N(0, 1) iff f Z (z) = 1 2π e z2 /2. Defn: Z R p MV N p (0, I) if and only if Z = (Z 1,..., Z p ) T with the Z i independent and each Z i N(0, 1). In this

More information

Nonlife Actuarial Models. Chapter 14 Basic Monte Carlo Methods

Nonlife Actuarial Models. Chapter 14 Basic Monte Carlo Methods Nonlife Actuarial Models Chapter 14 Basic Monte Carlo Methods Learning Objectives 1. Generation of uniform random numbers, mixed congruential method 2. Low discrepancy sequence 3. Inversion transformation

More information

Numerical Integration of SDEs: A Short Tutorial

Numerical Integration of SDEs: A Short Tutorial Numerical Integration of SDEs: A Short Tutorial Thomas Schaffter January 19, 010 1 Introduction 1.1 Itô and Stratonovich SDEs 1-dimensional stochastic differentiable equation (SDE) is given by [6, 7] dx

More information

1. Aufgabenblatt zur Vorlesung Probability Theory

1. Aufgabenblatt zur Vorlesung Probability Theory 24.10.17 1. Aufgabenblatt zur Vorlesung By (Ω, A, P ) we always enote the unerlying probability space, unless state otherwise. 1. Let r > 0, an efine f(x) = 1 [0, [ (x) exp( r x), x R. a) Show that p f

More information

An Introduction to Malliavin calculus and its applications

An Introduction to Malliavin calculus and its applications An Introduction to Malliavin calculus and its applications Lecture 3: Clark-Ocone formula David Nualart Department of Mathematics Kansas University University of Wyoming Summer School 214 David Nualart

More information

An Uncertain Control Model with Application to. Production-Inventory System

An Uncertain Control Model with Application to. Production-Inventory System An Uncertain Control Model with Application to Production-Inventory System Kai Yao 1, Zhongfeng Qin 2 1 Department of Mathematical Sciences, Tsinghua University, Beijing 100084, China 2 School of Economics

More information

Brownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539

Brownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539 Brownian motion Samy Tindel Purdue University Probability Theory 2 - MA 539 Mostly taken from Brownian Motion and Stochastic Calculus by I. Karatzas and S. Shreve Samy T. Brownian motion Probability Theory

More information

arxiv: v2 [math.na] 1 Aug 2016

arxiv: v2 [math.na] 1 Aug 2016 STOCHASTIC C-STABILITY AND B-CONSISTENCY OF EXPLICIT AND IMPLICIT MILSTEIN-TYPE SCHEMES WOLF-JÜRGEN BEYN, ELENA ISAAK, AND RAPHAEL KRUSE arxiv:151.06905v [math.na] 1 Aug 016 Abstract. This paper focuses

More information

Poisson random measure: motivation

Poisson random measure: motivation : motivation The Lévy measure provides the expected number of jumps by time unit, i.e. in a time interval of the form: [t, t + 1], and of a certain size Example: ν([1, )) is the expected number of jumps

More information

MATH4210 Financial Mathematics ( ) Tutorial 7

MATH4210 Financial Mathematics ( ) Tutorial 7 MATH40 Financial Mathematics (05-06) Tutorial 7 Review of some basic Probability: The triple (Ω, F, P) is called a probability space, where Ω denotes the sample space and F is the set of event (σ algebra

More information

B8.3 Mathematical Models for Financial Derivatives. Hilary Term Solution Sheet 2

B8.3 Mathematical Models for Financial Derivatives. Hilary Term Solution Sheet 2 B8.3 Mathematical Models for Financial Derivatives Hilary Term 18 Solution Sheet In the following W t ) t denotes a standard Brownian motion and t > denotes time. A partition π of the interval, t is a

More information

Simulation Method for Solving Stochastic Differential Equations with Constant Diffusion Coefficients

Simulation Method for Solving Stochastic Differential Equations with Constant Diffusion Coefficients Journal of mathematics and computer Science 8 (2014) 28-32 Simulation Method for Solving Stochastic Differential Equations with Constant Diffusion Coefficients Behrouz Fathi Vajargah Department of statistics,

More information

A Concise Course on Stochastic Partial Differential Equations

A Concise Course on Stochastic Partial Differential Equations A Concise Course on Stochastic Partial Differential Equations Michael Röckner Reference: C. Prevot, M. Röckner: Springer LN in Math. 1905, Berlin (2007) And see the references therein for the original

More information

2 Random Variable Generation

2 Random Variable Generation 2 Random Variable Generation Most Monte Carlo computations require, as a starting point, a sequence of i.i.d. random variables with given marginal distribution. We describe here some of the basic methods

More information

The concentration of a drug in blood. Exponential decay. Different realizations. Exponential decay with noise. dc(t) dt.

The concentration of a drug in blood. Exponential decay. Different realizations. Exponential decay with noise. dc(t) dt. The concentration of a drug in blood Exponential decay C12 concentration 2 4 6 8 1 C12 concentration 2 4 6 8 1 dc(t) dt = µc(t) C(t) = C()e µt 2 4 6 8 1 12 time in minutes 2 4 6 8 1 12 time in minutes

More information

FE 5204 Stochastic Differential Equations

FE 5204 Stochastic Differential Equations Instructor: Jim Zhu e-mail:zhu@wmich.edu http://homepages.wmich.edu/ zhu/ January 20, 2009 Preliminaries for dealing with continuous random processes. Brownian motions. Our main reference for this lecture

More information

( ) ( ) Monte Carlo Methods Interested in. E f X = f x d x. Examples:

( ) ( ) Monte Carlo Methods Interested in. E f X = f x d x. Examples: Monte Carlo Methods Interested in Examples: µ E f X = f x d x Type I error rate of a hypothesis test Mean width of a confidence interval procedure Evaluating a likelihood Finding posterior mean and variance

More information

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3 Brownian Motion Contents 1 Definition 2 1.1 Brownian Motion................................. 2 1.2 Wiener measure.................................. 3 2 Construction 4 2.1 Gaussian process.................................

More information

Stochastic Differential Equations

Stochastic Differential Equations Chapter 5 Stochastic Differential Equations We would like to introduce stochastic ODE s without going first through the machinery of stochastic integrals. 5.1 Itô Integrals and Itô Differential Equations

More information

Scientific Computing

Scientific Computing 2301678 Scientific Computing Chapter 2 Interpolation and Approximation Paisan Nakmahachalasint Paisan.N@chula.ac.th Chapter 2 Interpolation and Approximation p. 1/66 Contents 1. Polynomial interpolation

More information

= F (b) F (a) F (x i ) F (x i+1 ). a x 0 x 1 x n b i

= F (b) F (a) F (x i ) F (x i+1 ). a x 0 x 1 x n b i Real Analysis Problem 1. If F : R R is a monotone function, show that F T V ([a,b]) = F (b) F (a) for any interval [a, b], and that F has bounded variation on R if and only if it is bounded. Here F T V

More information

Malliavin Calculus in Finance

Malliavin Calculus in Finance Malliavin Calculus in Finance Peter K. Friz 1 Greeks and the logarithmic derivative trick Model an underlying assent by a Markov process with values in R m with dynamics described by the SDE dx t = b(x

More information

MA8109 Stochastic Processes in Systems Theory Autumn 2013

MA8109 Stochastic Processes in Systems Theory Autumn 2013 Norwegian University of Science and Technology Department of Mathematical Sciences MA819 Stochastic Processes in Systems Theory Autumn 213 1 MA819 Exam 23, problem 3b This is a linear equation of the form

More information

Mean-square Stability Analysis of an Extended Euler-Maruyama Method for a System of Stochastic Differential Equations

Mean-square Stability Analysis of an Extended Euler-Maruyama Method for a System of Stochastic Differential Equations Mean-square Stability Analysis of an Extended Euler-Maruyama Method for a System of Stochastic Differential Equations Ram Sharan Adhikari Assistant Professor Of Mathematics Rogers State University Mathematical

More information

Improved Discrepancy Bounds for Hybrid Sequences. Harald Niederreiter. RICAM Linz and University of Salzburg

Improved Discrepancy Bounds for Hybrid Sequences. Harald Niederreiter. RICAM Linz and University of Salzburg Improved Discrepancy Bounds for Hybrid Sequences Harald Niederreiter RICAM Linz and University of Salzburg MC vs. QMC methods Hybrid sequences The basic sequences Deterministic discrepancy bounds The proof

More information

Stochastic Differential Equations

Stochastic Differential Equations CHAPTER 1 Stochastic Differential Equations Consider a stochastic process X t satisfying dx t = bt, X t,w t dt + σt, X t,w t dw t. 1.1 Question. 1 Can we obtain the existence and uniqueness theorem for

More information

Scientific Computing: Monte Carlo

Scientific Computing: Monte Carlo Scientific Computing: Monte Carlo Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 Course MATH-GA.2043 or CSCI-GA.2112, Spring 2012 April 5th and 12th, 2012 A. Donev (Courant Institute)

More information

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games Alberto Bressan ) and Khai T. Nguyen ) *) Department of Mathematics, Penn State University **) Department of Mathematics,

More information

Exercises for Multivariable Differential Calculus XM521

Exercises for Multivariable Differential Calculus XM521 This document lists all the exercises for XM521. The Type I (True/False) exercises will be given, and should be answered, online immediately following each lecture. The Type III exercises are to be done

More information

I forgot to mention last time: in the Ito formula for two standard processes, putting

I forgot to mention last time: in the Ito formula for two standard processes, putting I forgot to mention last time: in the Ito formula for two standard processes, putting dx t = a t dt + b t db t dy t = α t dt + β t db t, and taking f(x, y = xy, one has f x = y, f y = x, and f xx = f yy

More information

THE INVERSE FUNCTION THEOREM

THE INVERSE FUNCTION THEOREM THE INVERSE FUNCTION THEOREM W. PATRICK HOOPER The implicit function theorem is the following result: Theorem 1. Let f be a C 1 function from a neighborhood of a point a R n into R n. Suppose A = Df(a)

More information

Stochastic Differential Equations.

Stochastic Differential Equations. Chapter 3 Stochastic Differential Equations. 3.1 Existence and Uniqueness. One of the ways of constructing a Diffusion process is to solve the stochastic differential equation dx(t) = σ(t, x(t)) dβ(t)

More information

Stochastic Integration and Stochastic Differential Equations: a gentle introduction

Stochastic Integration and Stochastic Differential Equations: a gentle introduction Stochastic Integration and Stochastic Differential Equations: a gentle introduction Oleg Makhnin New Mexico Tech Dept. of Mathematics October 26, 27 Intro: why Stochastic? Brownian Motion/ Wiener process

More information

Existence and Uniqueness

Existence and Uniqueness Chapter 3 Existence and Uniqueness An intellect which at a certain moment would know all forces that set nature in motion, and all positions of all items of which nature is composed, if this intellect

More information

conditional cdf, conditional pdf, total probability theorem?

conditional cdf, conditional pdf, total probability theorem? 6 Multiple Random Variables 6.0 INTRODUCTION scalar vs. random variable cdf, pdf transformation of a random variable conditional cdf, conditional pdf, total probability theorem expectation of a random

More information

The multidimensional Ito Integral and the multidimensional Ito Formula. Eric Mu ller June 1, 2015 Seminar on Stochastic Geometry and its applications

The multidimensional Ito Integral and the multidimensional Ito Formula. Eric Mu ller June 1, 2015 Seminar on Stochastic Geometry and its applications The multidimensional Ito Integral and the multidimensional Ito Formula Eric Mu ller June 1, 215 Seminar on Stochastic Geometry and its applications page 2 Seminar on Stochastic Geometry and its applications

More information

The Codimension of the Zeros of a Stable Process in Random Scenery

The Codimension of the Zeros of a Stable Process in Random Scenery The Codimension of the Zeros of a Stable Process in Random Scenery Davar Khoshnevisan The University of Utah, Department of Mathematics Salt Lake City, UT 84105 0090, U.S.A. davar@math.utah.edu http://www.math.utah.edu/~davar

More information

Bounded uniformly continuous functions

Bounded uniformly continuous functions Bounded uniformly continuous functions Objectives. To study the basic properties of the C -algebra of the bounded uniformly continuous functions on some metric space. Requirements. Basic concepts of analysis:

More information

Next tool is Partial ACF; mathematical tools first. The Multivariate Normal Distribution. e z2 /2. f Z (z) = 1 2π. e z2 i /2

Next tool is Partial ACF; mathematical tools first. The Multivariate Normal Distribution. e z2 /2. f Z (z) = 1 2π. e z2 i /2 Next tool is Partial ACF; mathematical tools first. The Multivariate Normal Distribution Defn: Z R 1 N(0,1) iff f Z (z) = 1 2π e z2 /2 Defn: Z R p MV N p (0, I) if and only if Z = (Z 1,..., Z p ) (a column

More information

Continuous Random Variables

Continuous Random Variables 1 / 24 Continuous Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 27, 2013 2 / 24 Continuous Random Variables

More information

MAS223 Statistical Inference and Modelling Exercises

MAS223 Statistical Inference and Modelling Exercises MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,

More information

Covariance. Lecture 20: Covariance / Correlation & General Bivariate Normal. Covariance, cont. Properties of Covariance

Covariance. Lecture 20: Covariance / Correlation & General Bivariate Normal. Covariance, cont. Properties of Covariance Covariance Lecture 0: Covariance / Correlation & General Bivariate Normal Sta30 / Mth 30 We have previously discussed Covariance in relation to the variance of the sum of two random variables Review Lecture

More information

4 Uniform convergence

4 Uniform convergence 4 Uniform convergence In the last few sections we have seen several functions which have been defined via series or integrals. We now want to develop tools that will allow us to show that these functions

More information

Introduction to the Numerical Solution of SDEs Selected topics

Introduction to the Numerical Solution of SDEs Selected topics 0 / 23 Introduction to the Numerical Solution of SDEs Selected topics Andreas Rößler Summer School on Numerical Methods for Stochastic Differential Equations at Vienna University of Technology, Austria

More information

The Multivariate Gaussian Distribution

The Multivariate Gaussian Distribution The Multivariate Gaussian Distribution Chuong B. Do October, 8 A vector-valued random variable X = T X X n is said to have a multivariate normal or Gaussian) distribution with mean µ R n and covariance

More information

The moment-generating function of the log-normal distribution using the star probability measure

The moment-generating function of the log-normal distribution using the star probability measure Noname manuscript No. (will be inserted by the editor) The moment-generating function of the log-normal distribution using the star probability measure Yuri Heymann Received: date / Accepted: date Abstract

More information

The Hilbert Space of Random Variables

The Hilbert Space of Random Variables The Hilbert Space of Random Variables Electrical Engineering 126 (UC Berkeley) Spring 2018 1 Outline Fix a probability space and consider the set H := {X : X is a real-valued random variable with E[X 2

More information

Stochastic Numerical Analysis

Stochastic Numerical Analysis Stochastic Numerical Analysis Prof. Mike Giles mike.giles@maths.ox.ac.uk Oxford University Mathematical Institute Stoch. NA, Lecture 3 p. 1 Multi-dimensional SDEs So far we have considered scalar SDEs

More information

Random Variables and Their Distributions

Random Variables and Their Distributions Chapter 3 Random Variables and Their Distributions A random variable (r.v.) is a function that assigns one and only one numerical value to each simple event in an experiment. We will denote r.vs by capital

More information

Explicit Bounds for the Distribution Function of the Sum of Dependent Normally Distributed Random Variables

Explicit Bounds for the Distribution Function of the Sum of Dependent Normally Distributed Random Variables Explicit Bounds for the Distribution Function of the Sum of Dependent Normally Distributed Random Variables Walter Schneider July 26, 20 Abstract In this paper an analytic expression is given for the bounds

More information

Random processes and probability distributions. Phys 420/580 Lecture 20

Random processes and probability distributions. Phys 420/580 Lecture 20 Random processes and probability distributions Phys 420/580 Lecture 20 Random processes Many physical processes are random in character: e.g., nuclear decay (Poisson distributed event count) P (k, τ) =

More information

The Metropolis-Hastings Algorithm. June 8, 2012

The Metropolis-Hastings Algorithm. June 8, 2012 The Metropolis-Hastings Algorithm June 8, 22 The Plan. Understand what a simulated distribution is 2. Understand why the Metropolis-Hastings algorithm works 3. Learn how to apply the Metropolis-Hastings

More information

For a stochastic process {Y t : t = 0, ±1, ±2, ±3, }, the mean function is defined by (2.2.1) ± 2..., γ t,

For a stochastic process {Y t : t = 0, ±1, ±2, ±3, }, the mean function is defined by (2.2.1) ± 2..., γ t, CHAPTER 2 FUNDAMENTAL CONCEPTS This chapter describes the fundamental concepts in the theory of time series models. In particular, we introduce the concepts of stochastic processes, mean and covariance

More information

ISyE 6644 Fall 2014 Test 3 Solutions

ISyE 6644 Fall 2014 Test 3 Solutions 1 NAME ISyE 6644 Fall 14 Test 3 Solutions revised 8/4/18 You have 1 minutes for this test. You are allowed three cheat sheets. Circle all final answers. Good luck! 1. [4 points] Suppose that the joint

More information

Numerical Methods for Mathematical Finance

Numerical Methods for Mathematical Finance Numerical Methods for Mathematical Finance Peter Philip Lecture Notes Created for the Class of Spring Semester 21 at LMU Munich May 3, 211 Contents 1 Introduction: Review and Motivation 4 1.1 Review: Probability

More information

Course: ESO-209 Home Work: 1 Instructor: Debasis Kundu

Course: ESO-209 Home Work: 1 Instructor: Debasis Kundu Home Work: 1 1. Describe the sample space when a coin is tossed (a) once, (b) three times, (c) n times, (d) an infinite number of times. 2. A coin is tossed until for the first time the same result appear

More information

JUSTIN HARTMANN. F n Σ.

JUSTIN HARTMANN. F n Σ. BROWNIAN MOTION JUSTIN HARTMANN Abstract. This paper begins to explore a rigorous introduction to probability theory using ideas from algebra, measure theory, and other areas. We start with a basic explanation

More information

Stochastic Simulation Variance reduction methods Bo Friis Nielsen

Stochastic Simulation Variance reduction methods Bo Friis Nielsen Stochastic Simulation Variance reduction methods Bo Friis Nielsen Applied Mathematics and Computer Science Technical University of Denmark 2800 Kgs. Lyngby Denmark Email: bfni@dtu.dk Variance reduction

More information

Wiener Measure and Brownian Motion

Wiener Measure and Brownian Motion Chapter 16 Wiener Measure and Brownian Motion Diffusion of particles is a product of their apparently random motion. The density u(t, x) of diffusing particles satisfies the diffusion equation (16.1) u

More information

E[X n ]= dn dt n M X(t). ). What is the mgf? Solution. Found this the other day in the Kernel matching exercise: 1 M X (t) =

E[X n ]= dn dt n M X(t). ). What is the mgf? Solution. Found this the other day in the Kernel matching exercise: 1 M X (t) = Chapter 7 Generating functions Definition 7.. Let X be a random variable. The moment generating function is given by M X (t) =E[e tx ], provided that the expectation exists for t in some neighborhood of

More information

DUBLIN CITY UNIVERSITY

DUBLIN CITY UNIVERSITY DUBLIN CITY UNIVERSITY SAMPLE EXAMINATIONS 2017/2018 MODULE: QUALIFICATIONS: Simulation for Finance MS455 B.Sc. Actuarial Mathematics ACM B.Sc. Financial Mathematics FIM YEAR OF STUDY: 4 EXAMINERS: Mr

More information

University of Regina. Lecture Notes. Michael Kozdron

University of Regina. Lecture Notes. Michael Kozdron University of Regina Statistics 252 Mathematical Statistics Lecture Notes Winter 2005 Michael Kozdron kozdron@math.uregina.ca www.math.uregina.ca/ kozdron Contents 1 The Basic Idea of Statistics: Estimating

More information

Verona Course April Lecture 1. Review of probability

Verona Course April Lecture 1. Review of probability Verona Course April 215. Lecture 1. Review of probability Viorel Barbu Al.I. Cuza University of Iaşi and the Romanian Academy A probability space is a triple (Ω, F, P) where Ω is an abstract set, F is

More information

Gaussian vectors and central limit theorem

Gaussian vectors and central limit theorem Gaussian vectors and central limit theorem Samy Tindel Purdue University Probability Theory 2 - MA 539 Samy T. Gaussian vectors & CLT Probability Theory 1 / 86 Outline 1 Real Gaussian random variables

More information

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015 Part IA Probability Theorems Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.

More information

ISyE 6644 Fall 2014 Test #2 Solutions (revised 11/7/16)

ISyE 6644 Fall 2014 Test #2 Solutions (revised 11/7/16) 1 NAME ISyE 6644 Fall 2014 Test #2 Solutions (revised 11/7/16) This test is 85 minutes. You are allowed two cheat sheets. Good luck! 1. Some short answer questions to get things going. (a) Consider the

More information

Chapter 2 Generating Random Numbers with Specified Distributions

Chapter 2 Generating Random Numbers with Specified Distributions Chapter 2 Generating Random Numbers with Specified Distributions Simulation and valuation of finance instruments require numbers with specified distributions. For example, in Section.6 we have used numbers

More information

S6880 #13. Variance Reduction Methods

S6880 #13. Variance Reduction Methods S6880 #13 Variance Reduction Methods 1 Variance Reduction Methods Variance Reduction Methods 2 Importance Sampling Importance Sampling 3 Control Variates Control Variates Cauchy Example Revisited 4 Antithetic

More information

A MODEL FOR THE LONG-TERM OPTIMAL CAPACITY LEVEL OF AN INVESTMENT PROJECT

A MODEL FOR THE LONG-TERM OPTIMAL CAPACITY LEVEL OF AN INVESTMENT PROJECT A MODEL FOR HE LONG-ERM OPIMAL CAPACIY LEVEL OF AN INVESMEN PROJEC ARNE LØKKA AND MIHAIL ZERVOS Abstract. We consider an investment project that produces a single commodity. he project s operation yields

More information

Phenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 2012

Phenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 2012 Phenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 202 BOUNDS AND ASYMPTOTICS FOR FISHER INFORMATION IN THE CENTRAL LIMIT THEOREM

More information

Multivariate Random Variable

Multivariate Random Variable Multivariate Random Variable Author: Author: Andrés Hincapié and Linyi Cao This Version: August 7, 2016 Multivariate Random Variable 3 Now we consider models with more than one r.v. These are called multivariate

More information

Exercises in stochastic analysis

Exercises in stochastic analysis Exercises in stochastic analysis Franco Flandoli, Mario Maurelli, Dario Trevisan The exercises with a P are those which have been done totally or partially) in the previous lectures; the exercises with

More information

Hochdimensionale Integration

Hochdimensionale Integration Oliver Ernst Institut für Numerische Mathematik und Optimierung Hochdimensionale Integration 14-tägige Vorlesung im Wintersemester 2010/11 im Rahmen des Moduls Ausgewählte Kapitel der Numerik Contents

More information

the convolution of f and g) given by

the convolution of f and g) given by 09:53 /5/2000 TOPIC Characteristic functions, cont d This lecture develops an inversion formula for recovering the density of a smooth random variable X from its characteristic function, and uses that

More information

MTH739U/P: Topics in Scientific Computing Autumn 2016 Week 6

MTH739U/P: Topics in Scientific Computing Autumn 2016 Week 6 MTH739U/P: Topics in Scientific Computing Autumn 16 Week 6 4.5 Generic algorithms for non-uniform variates We have seen that sampling from a uniform distribution in [, 1] is a relatively straightforward

More information

18 Bivariate normal distribution I

18 Bivariate normal distribution I 8 Bivariate normal distribution I 8 Example Imagine firing arrows at a target Hopefully they will fall close to the target centre As we fire more arrows we find a high density near the centre and fewer

More information

Lecture 2. We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales.

Lecture 2. We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales. Lecture 2 1 Martingales We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales. 1.1 Doob s inequality We have the following maximal

More information

for all subintervals I J. If the same is true for the dyadic subintervals I D J only, we will write ϕ BMO d (J). In fact, the following is true

for all subintervals I J. If the same is true for the dyadic subintervals I D J only, we will write ϕ BMO d (J). In fact, the following is true 3 ohn Nirenberg inequality, Part I A function ϕ L () belongs to the space BMO() if sup ϕ(s) ϕ I I I < for all subintervals I If the same is true for the dyadic subintervals I D only, we will write ϕ BMO

More information

WEAK VERSIONS OF STOCHASTIC ADAMS-BASHFORTH AND SEMI-IMPLICIT LEAPFROG SCHEMES FOR SDES. 1. Introduction

WEAK VERSIONS OF STOCHASTIC ADAMS-BASHFORTH AND SEMI-IMPLICIT LEAPFROG SCHEMES FOR SDES. 1. Introduction WEAK VERSIONS OF STOCHASTIC ADAMS-BASHFORTH AND SEMI-IMPLICIT LEAPFROG SCHEMES FOR SDES BRIAN D. EWALD 1 Abstract. We consider the weak analogues of certain strong stochastic numerical schemes considered

More information

P (A G) dp G P (A G)

P (A G) dp G P (A G) First homework assignment. Due at 12:15 on 22 September 2016. Homework 1. We roll two dices. X is the result of one of them and Z the sum of the results. Find E [X Z. Homework 2. Let X be a r.v.. Assume

More information

Probability and Measure

Probability and Measure Part II Year 2018 2017 2016 2015 2014 2013 2012 2011 2010 2009 2008 2007 2006 2005 2018 84 Paper 4, Section II 26J Let (X, A) be a measurable space. Let T : X X be a measurable map, and µ a probability

More information