TMS165/MSA350 Stochastic Calculus, Lecture on Applications

Similar documents
Kolmogorov Equations and Markov Processes

Lecture 21: Stochastic Differential Equations

p 1 ( Y p dp) 1/p ( X p dp) 1 1 p

Lecture 12: Diffusion Processes and Stochastic Differential Equations

Introduction to Diffusion Processes.

I forgot to mention last time: in the Ito formula for two standard processes, putting

Homogenization with stochastic differential equations

Stochastic Calculus and Black-Scholes Theory MTH772P Exercises Sheet 1

Introduction to multiscale modeling and simulation. Explicit methods for ODEs : forward Euler. y n+1 = y n + tf(y n ) dy dt = f(y), y(0) = y 0

Sloppy derivations of Ito s formula and the Fokker-Planck equations

STATISTICS 385: STOCHASTIC CALCULUS HOMEWORK ASSIGNMENT 4 DUE NOVEMBER 23, = (2n 1)(2n 3) 3 1.

Lecture 6: Bayesian Inference in SDE Models

Introduction to numerical simulations for Stochastic ODEs

The multidimensional Ito Integral and the multidimensional Ito Formula. Eric Mu ller June 1, 2015 Seminar on Stochastic Geometry and its applications

Lecture 1: Pragmatic Introduction to Stochastic Differential Equations

Stochastic differential equation models in biology Susanne Ditlevsen

{σ x >t}p x. (σ x >t)=e at.

Stochastic Calculus February 11, / 33

LAN property for ergodic jump-diffusion processes with discrete observations

Convoluted Brownian motions: a class of remarkable Gaussian processes

Research Article A Necessary Characteristic Equation of Diffusion Processes Having Gaussian Marginals

The concentration of a drug in blood. Exponential decay. Different realizations. Exponential decay with noise. dc(t) dt.

Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics. Jiti Gao

Handbook of Stochastic Methods

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 7 9/25/2013

Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( )

SDE Coefficients. March 4, 2008

Lecture on Parameter Estimation for Stochastic Differential Equations. Erik Lindström

Lecture 4: Introduction to stochastic processes and stochastic calculus

Bernardo D Auria Stochastic Processes /10. Notes. Abril 13 th, 2010

Lecture 4: Ito s Stochastic Calculus and SDE. Seung Yeal Ha Dept of Mathematical Sciences Seoul National University

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

Approximating diffusions by piecewise constant parameters

6. Brownian Motion. Q(A) = P [ ω : x(, ω) A )

Exercises. T 2T. e ita φ(t)dt.

The Smoluchowski-Kramers Approximation: What model describes a Brownian particle?

Backward Stochastic Differential Equations with Infinite Time Horizon

Stochastic Integration.

Poisson Jumps in Credit Risk Modeling: a Partial Integro-differential Equation Formulation

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

Stochastic Differential Equations.

Lecture 12: Detailed balance and Eigenfunction methods

A Short Introduction to Diffusion Processes and Ito Calculus

M4A42 APPLIED STOCHASTIC PROCESSES

Theoretical Tutorial Session 2

Bernardo D Auria Stochastic Processes /12. Notes. March 29 th, 2012

Växjö University. Some recent simulation techniques for diffusion bridges. School of Mathematics and System Engineering

Introduction. Stochastic Processes. Will Penny. Stochastic Differential Equations. Stochastic Chain Rule. Expectations.

Affine Processes. Econometric specifications. Eduardo Rossi. University of Pavia. March 17, 2009

Stochastic Differential Equations

Estimation of arrival and service rates for M/M/c queue system

Brownian Motion and Stochastic Calculus

2 Quick Review of Continuous Random Variables

Introduction to Probability and Stocastic Processes - Part I

Solution for Problem 7.1. We argue by contradiction. If the limit were not infinite, then since τ M (ω) is nondecreasing we would have

Week 9 Generators, duality, change of measure

Brownian Motion. An Undergraduate Introduction to Financial Mathematics. J. Robert Buchanan. J. Robert Buchanan Brownian Motion

The Cameron-Martin-Girsanov (CMG) Theorem

Interest Rate Models:

Short-time expansions for close-to-the-money options under a Lévy jump model with stochastic volatility

Stochastic optimal control with rough paths

This is a Gaussian probability centered around m = 0 (the most probable and mean position is the origin) and the mean square displacement m 2 = n,or

Quantization of stochastic processes with applications on Euler-Maruyama schemes

A new approach for investment performance measurement. 3rd WCMF, Santa Barbara November 2009

MA8109 Stochastic Processes in Systems Theory Autumn 2013

Multivariate Generalized Ornstein-Uhlenbeck Processes

Stochastic process. X, a series of random variables indexed by t

1 Simulating normal (Gaussian) rvs with applications to simulating Brownian motion and geometric Brownian motion in one and two dimensions

Lecture 4: Numerical Solution of SDEs, Itô Taylor Series, Gaussian Approximations

Lecture 6: Multiple Model Filtering, Particle Filtering and Other Approximations

An adaptive numerical scheme for fractional differential equations with explosions

Handbook of Stochastic Methods

FIRST-ORDER ORDINARY DIFFERENTIAL EQUATIONS I: Introduction and Analytic Methods David Levermore Department of Mathematics University of Maryland

Least Squares Estimators for Stochastic Differential Equations Driven by Small Lévy Noises

Stationary distributions of non Gaussian Ornstein Uhlenbeck processes for beam halos

Nonparametric Drift Estimation for Stochastic Differential Equations

Birth and Death Processes. Birth and Death Processes. Linear Growth with Immigration. Limiting Behaviour for Birth and Death Processes

Gaussian processes for inference in stochastic differential equations

Local vs. Nonlocal Diffusions A Tale of Two Laplacians

Infinite-dimensional methods for path-dependent equations

Lecture 4: Numerical Solution of SDEs, Itô Taylor Series, Gaussian Process Approximations

DISCRETE-TIME STOCHASTIC MODELS, SDEs, AND NUMERICAL METHODS. Ed Allen. NIMBioS Tutorial: Stochastic Models With Biological Applications

Brownian Motion. Chapter Stochastic Process

6 Continuous-Time Birth and Death Chains

ELEMENTS OF PROBABILITY THEORY

Exercise Sheet 7 - Solutions

Weak solutions of mean-field stochastic differential equations

Likelihood Functions for Stochastic Signals in White Noise* TYRONE E. DUNCAN

Rough paths methods 1: Introduction

Exact Simulation of Diffusions and Jump Diffusions

Lecture 12: Detailed balance and Eigenfunction methods

Continuous-Time Markov Chain

1. Stochastic Process

A Concise Course on Stochastic Partial Differential Equations

On the Rate of Convergence to Equilibrium for Two-sided Reflected Brownian Motion and for the Ornstein-Uhlenbeck Process

Chain Rule. MATH 311, Calculus III. J. Robert Buchanan. Spring Department of Mathematics

Stochastic Modelling Unit 1: Markov chain models

Deterministic Dynamic Programming

Asymptotic properties of maximum likelihood estimator for the growth rate for a jump-type CIR process

Path integrals for classical Markov processes

Transcription:

TMS165/MSA35 Stochastic Calculus, Lecture on Applications In this lecture we demonstrate how statistical methods such as the maximum likelihood method likelihood ratio estimation can be applied to the Ornstein-Uhlenbeck (OU) process. 1. Elements of diffusion theory Diffusion processes SDE A time homogeneous diffusion process is the solution X = {X(t) t to an SDE of the form dx(t) = µ(x(t))dt+σ(x(t))db(t) for t, (1) where the drift coefficient µ : the diffusion coefficient σ : are sufficiently nice functions. Here B = {B(t) t denotes a Brownian motion as usual. By definition, the solution X to (1) satisfies X(t) = X()+ Markov property t µ(x(r))dr + t σ(x(r))db(r) for t. () The solution X to the SDE (1) is a Markov process, which is to say that P { X(t) F X s = P{X(t) X(s) for s t. Here Fs X = σ{x(r):r s for s is the filtration generated by the process X itself. While a detailed rigorous proof of the Markov property is extremly complicated, the Markov property is easy to underst from a more heursitic point of view: Using the representation () for both X(t) X(s) we get X(t) = X(s)+ t s = X(s)+lim µ(x(r))dr + t s σ(x(r)) db(r) n µ(x(t i 1 ))(t i t i 1 )+lim n σ(x(t i 1 ))(B(t i ) B(t i 1 )), where s = t < t 1 <... < t n = t is a partion of the interval [s,t] that becomes finer finer in the limit. From this we see the that the only thing from the past Fs X that affects the future value X(t) is the value X(s)=X(t ) of X at time s=t. 1

Transition densities The transition density function p(t,x,y) = d P{X(t+s) y X(s)=x for t> dy of the diffusion process X given by (1) or () satisfies the Kolmogorov backward PDE σ(x) p(t,x,y) = t x p(t,x,y)+µ(x) x p(t,x,y). Conversely, under general conditions, a solution to this PDE is the transition density function of the diffusion process given by (1) or () if it is a probability density function as a function of y for each choice of x t>, that is, p(t,x,y) p(t,x,y)dy = 1, in addition satisfies p(t,x,y) as t for x y. In general it is not an easy task or even possible to find an explicit expression for the transition density function. Argubly, the most common systematic way to try to solve the Kolmogorov backward PDE is to consider the Laplace transform ˆp(λ,x,y) = which must satisfy the ODE (for x y), as λ ˆp(λ,x,y) = σ(x) e λt p(t,x,y)dt for λ>, x ˆp(λ,x,y)+µ(x) x ˆp(λ,x,y) e λt t p(t,x,y)dt = λ e λt p(t,x,y)dt (for x y) when p(,x,y)=. The conditions that p is a density function translates to ˆp(λ,x,y)dy = e λt p(t,x,y)dtdy = e λt p(t,x,y)dydt = e λt dt = 1 λ. As the ODE for ˆp(λ,x,y) usually has a unique solution that integrates to 1/λ, in the above fashion, this determines the Laplace transform ˆp(λ, x, y) of p(t, x, y), after which p(t, x, y) is found by inverse Laplace tranformation. Although the above indicated way to find transition densities for diffusions does indeed work for many important equations, the details of the solution are in general too difficult to be attempted on undergraduate level, it is usually more rewarding to search the literature (web) for solutions than to try to derive them oneself.

Stationary distribution A stationary density function for X is a probability density function π that satisfies d π(y) = dy P{X(t+s) y X(s)=xπ(x)dx = p(t,x,y)π(x)dx for t>. By the Chapman-Kolmogorov equation below Theorem 5.6 in Klebaner s book, this means that if the process has the stationary density function (distribution) at a certain time, then it has the stationary density function (distribution) at all later times. It can be shown (see Task of Home exercise session 5) that if X is started according to the stationary distribution at time t =, then X is a stationary process. Also, regardless of how X is started, X will converge (in a way we select not to specify) to a stationary process with the stationary marginal distribution as t. As the transition density satisfies the Kolmogorov forward PDE ( ) σ(y) p(t,x,y) = t y p(t,x,y) ( ) µ(y)p(t,x,y), y we see from the above integral equation that ( σ(y) y ( )π(y) y µ(y) = t σ(y) y + )π(y) y µ(y) ( = t σ(y) y + )p(t,x,y)π(x)dx y µ(y) =. If it exists, then the stationary density function is given by π(y) = C { y σ(y) exp µ(z) y σ(z) dz. Here C > is a normalizing constant selected to make π a probability density function, that is, π(y)dy = 1, while y is any constant. The existence of the stationary density function is the same thing as that this normalizing procedure is possible to carry out. Note that it is easy to see that π satisfies the ODE (cf. above) ( σ(y) y )π(y). y µ(y) Finite dimensional distributions The joint probability density function of (X(t ),X(t 1 ),...,X(t n )) for = t < t 1 <... <t n is given by f X(t ),X(t 1 ),...,X(t n)(x,x 1,...,x n ) = π(x ) 3 n p(t i t i 1,x i 1,x i )

when the process is started with the stationary density function π (provided that it exists) at time, while n f X(t1 ),...,X(t n)(x 1,...,x n ) = p(t i t i 1,x i 1,x i ) when the process is started at a fixed value X()=x. Euler method We may simulate a weak solution to the SDE at a time grid =t <t 1 <... <t n =T by means of the Euler method, as X(t i ) X(t i 1 )+µ(x(t i 1 ))(t i t i 1 )+σ(x(t i 1 ))ξ i for i = 1,...,n, where {ξ i n are independent rom variables such that ξ i is zero-mean normal distributed with starddeviation t i t i 1, where X() is a rom variable that is indepedent of {ξ i n has the stationary distribution if X is started according to the stationary distribution, while X()=x if X is started at a fixed value x. Likelihood ratios If the diffusion process X satisfies the equation for a P 1 -Brownian motion B dx(t) = µ 1 (X(t))dt+σ(X(t))dB(t) dx(t) = µ (X(t))dt+σ(X(t))dW(t) for a P -Brownian motion W, where P 1 P are two different probability measures, then the likelihood ratio between P P 1 based on an observation {X(t) t [,T] of the process X in the time interval [,T] is given by { dp T µ (X(t)) µ 1 (X(t)) = exp dp 1 σ(x(t)) dx(t) 1 T µ (X(t)) µ 1 (X(t)) σ(x(t)) dt. This likelihood ratio can be used to judge which is the most likely of the above two SDE-models for X based on the observations: If dp /dp 1 is bigger than 1 then the model with the drift µ is the most likely, while a value of dp /dp 1 that is smaller than 1 indicates that the drift µ 1 is the most likely. The likelihood ratio can also be used to estimate parameters of a parametric SDE, as we will see later. 4

. The OU process An OU process is the solution X to the SDE dx(t) = µx(t)dt+σdb(t), where µ> σ> are parameters. In other words, the drift is µ(x) = µx the difffusion coefficient σ(x) = σ. The OU process has stationary distribution π(y) = 1 y /( { { σ exp µz σ dz 1 σ exp y ) µz σ dz dy = µ πσ exp that is, a zero-mean normal distribution with starddeviation σ/ µ. { µy σ, As we have shown in Exercise Session 5, the OU process has transition density p(t,x,y) = exp (y e µt x) πσ 1 e µt σ (1 e µt, )/µ thatis,anormaldistributionwithmeane µt xstarddeviationσ 1 e µt / µ. It is also a routine matter to differentiate to check that this function p satisfies the Kolmogorov backward PDE. If X is an OU process dx(t) = µ 1 X(t)dt+σdB(t) for a P µ1 -Brownian motion B, an OU process dx(t) = µ X(t)dt+σdW(t) for a P µ -Brownian motion W, then the corresponding likelihood ratio is given by { dp µ = exp µ T T µ 1 dp µ1 σ X(t)dX(t) µ µ 1 σ X(t) dt. In particular, we can find which is the most likely of the models dx(t) = db(t) dx(t) = X(t)dt+dW(t) by computing the likelihood ratio for µ 1 =, µ =1 σ=1 { dp T 1 = exp X(t)dX(t) 1 T X(t) dt, dp 5

then check whether dp 1 /dp > 1, indicating that µ = 1 is the most appropriate model, or dp 1 /dp < 1, indicating that µ= is the most appropriate model. We can also estimate the parameter µ for the equation dx(t) = µx(t)dt+dw(t) by means of maximizing the likelihood { dp µ = exp µ dp T which by differentiation gives the estimate µ = T 3. Application to the OU process X(t)dX(t) µ T /( T ) X(t) dx(t) X(t) dt. X(t) dt, We used the Euler method to simulate an OU process {X(t) t [,1] started according to the stationary distribution, an OU process {Y(t) t [,1] started at zero. In both cases the drift was µ(x) = µ x the diffusion coefficient σ(x) = σ, where µ = σ = 1. We use distance 1 1 between the time points of the simulation grid, so that =t < t 1 <... <t 1 =1, where t i t i 1 = 1 1 for i = 1,...,1. The simulations were carried out by means of the following Mathematica programs. <<Statistics ContinuousDistributions ; dt=1/1; T=1; {mu,sigma={1,1; For[i=; X={om[NormalDistribution[,sigma/Sqrt[*mu]]], i<=t/dt, i++, AppendTo[X, X[[i-1]] - mu*x[[i-1]]*dt + om[normaldistribution[,sigma*sqrt[dt]]]]] For[i=; Y=, i<=t/dt, i++, AppendTo[Y, Y[[i-1]] - mu*y[[i-1]]*dt + om[normaldistribution[,sigma*sqrt[dt]]]]] The results of the simulations are depicted in the following two figures 6

1 4 6 8 1-1 Y t OU process dy t Y t dt db t, Y 1.5 1.5 -.5 4 6 8 1 t -1-1.5 The joint density functions of (X(t ),...,X(t 1 )) (Y(t ),...,Y(t 1 )) are given by f X(t ),...,X(t 1 )(x,...,x 1 ) 1 = π(x ) p(t i t i 1,x i 1,x i ) = µ πσ exp { 1 µx σ f Y(t1 ),...,Y(t 1 )(y 1,...,y 1 ) = = 1 1 exp (x i e µ/1 x i 1 ) πσ 1 e µ/1 σ (1 e µ/1 )/µ p(t i t i 1,y i 1,y i ) exp (y i e µ/1 y i 1 ) πσ 1 e µ/1 σ (1 e µ/1, )/µ respectively, where y =. Hence we may use the maximum likelihood method to estimate µ σ from our simulated data (pretending that they are unknown), by 7

means of maximizing the likelihood µ πσ exp { µx(t ) 1 σ 1 exp (X(t i) e µ/1 X(t i 1 )) πσ 1 e µ/1 σ (1 e µ/1 )/µ exp (Y(t i) e µ/1 Y(t i 1 )) πσ 1 e µ/1 σ (1 e µ/1, )/µ respectively. These maximum likelihood estimates were carried out by means of the following Mathematcia code (with the densities logged to not get numerical underflows). foustationary[mu,sigma,x ] := Sqrt[mu]*Exp[-mu*x^/sigma^]/(Sqrt[Pi]*sigma); pou[mu,sigma,x,y,t ] := Exp[-(y-x*Exp[-mu*t])^/(*sigma^*(1-Exp[-*mu*t])/(*mu))] /(Sqrt[*Pi]*sigma*Sqrt[1-Exp[-*mu*t]]/Sqrt[*mu]); MLStationary[mu,sigma,dt,Data ] := Log[fOUStationary[mu,sigma,Data[[1]]]] + Sum[Log[pOU[mu,sigma,Data[[i-1]],Data[[i]],dt]], {i,,length[data]] MLNonStationary[mu,sigma,dt,Data ] := Sum[Log[pOU[mu,sigma,Data[[i-1]],Data[[i]],dt]], {i,,length[data]] NMaximize[MLStationary[mu,sigma,dt,X],mu>,sigma>, mu,sigma] Out[]:= {89.45, {mu -> 1.1156, sigma ->.99633 NMaximize[MLNonStationary[mu,sigma,dt,Y],mu>,sigma>, mu,sigma] Out[]:= {886.86, {mu ->.99798, sigma -> 1.88 Note how well this fits with the correct values µ=1 σ=1 for the parameters. We calculate the likelihood ratios { dp 1 (X) = exp dp 1 X(t)dX(t) 1 1 X(t) dt { dp 1 (Y) = exp dp 1 Y(t)dY(t) 1 1 Y(t) dt, 8

for our simulated processes, in order to fine whether dx(t) = db(t) or dx(t) = X(t)dt+dW(t) dy(t) = db(t) or dy(t) = Y(t)dt+dW(t), respectively, are the most likely models for the data. As both ratios where significantly larger than 1 (see the enclosed Mathematica code), the model with µ=1 was the most likely for both data sets. OUatioTest[Data ] := Exp[Sum[-Data[[i-1]]*(Data[[i]]-Data[[i-1]]), {i,, Length[Data]] - Sum[Data[[i]]^*dt, {i,1,length[data]]/]; {OUatioTest[X], OUatioTest[Y] Out[]:= {1.134, 1.5346 We may estimate the parameter µ by means of maximizing the likelihood ratios (dp µ /dp )(X) (dp µ /dp )(Y), respectively, which gives the µ estimates µ = 1 /( 1 ) X(t) dx(t) X(t) dt µ = 1 /( 1 ) Y(t)dY(t) Y(t) dt, respectively. Both results were very close to the correct µ=1, as the following Mathematica code illustrates: OUatioEst[Data ] := -Sum[Data[[i-1]]*(Data[[i]]-Data[[i-1]]), {i,,length[data]] /Sum[Data[[i]]^*dt, {i,1,length[data]]; {OUatioEst[X], OUatioEst[Y] {.954688,.9917 9