(B(t i+1 ) B(t i )) 2

Similar documents
Stochastic Calculus (Lecture #3)

On pathwise stochastic integration

n E(X t T n = lim X s Tn = X s

I forgot to mention last time: in the Ito formula for two standard processes, putting

Lecture 19 L 2 -Stochastic integration

Elementary properties of functions with one-sided limits

MA8109 Stochastic Processes in Systems Theory Autumn 2013

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3

Review of Multi-Calculus (Study Guide for Spivak s CHAPTER ONE TO THREE)

An essay on the general theory of stochastic processes

Filtrations, Markov Processes and Martingales. Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition

1. Stochastic Processes and filtrations

16.1. Signal Process Observation Process The Filtering Problem Change of Measure

Stochastic Processes II/ Wahrscheinlichkeitstheorie III. Lecture Notes

Lecture 9. d N(0, 1). Now we fix n and think of a SRW on [0,1]. We take the k th step at time k n. and our increments are ± 1

Stochastic Differential Equations.

Analysis Finite and Infinite Sets The Real Numbers The Cantor Set

Bernardo D Auria Stochastic Processes /12. Notes. March 29 th, 2012

Topics in fractional Brownian motion

A NOTE ON STOCHASTIC INTEGRALS AS L 2 -CURVES

Exercises. T 2T. e ita φ(t)dt.

Lecture 12. F o s, (1.1) F t := s>t

Applications of Ito s Formula

The strictly 1/2-stable example

Stochastic Calculus. Kevin Sinclair. August 2, 2016

Measure and integration

STAT 7032 Probability Spring Wlodek Bryc

Functional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals

+ 2x sin x. f(b i ) f(a i ) < ɛ. i=1. i=1

I. ANALYSIS; PROBABILITY

d(x n, x) d(x n, x nk ) + d(x nk, x) where we chose any fixed k > N

Bernardo D Auria Stochastic Processes /10. Notes. Abril 13 th, 2010

FE610 Stochastic Calculus for Financial Engineers. Stevens Institute of Technology

n [ F (b j ) F (a j ) ], n j=1(a j, b j ] E (4.1)

REVIEW OF ESSENTIAL MATH 346 TOPICS

Stochastic Analysis I S.Kotani April 2006

A Concise Course on Stochastic Partial Differential Equations

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

Economics 204 Fall 2011 Problem Set 2 Suggested Solutions

Part III Stochastic Calculus and Applications

Solution for Problem 7.1. We argue by contradiction. If the limit were not infinite, then since τ M (ω) is nondecreasing we would have

3 (Due ). Let A X consist of points (x, y) such that either x or y is a rational number. Is A measurable? What is its Lebesgue measure?

Lecture 21 Representations of Martingales

MAT 570 REAL ANALYSIS LECTURE NOTES. Contents. 1. Sets Functions Countability Axiom of choice Equivalence relations 9

On Integration-by-parts and the Itô Formula for Backwards Itô Integral

GENERALIZED COVARIATION FOR BANACH SPACE VALUED PROCESSES, ITÔ FORMULA AND APPLICATIONS

Some SDEs with distributional drift Part I : General calculus. Flandoli, Franco; Russo, Francesco; Wolf, Jochen

STOCHASTIC CALCULUS JASON MILLER AND VITTORIA SILVESTRI

{σ x >t}p x. (σ x >t)=e at.

A Short Introduction to Diffusion Processes and Ito Calculus

Hardy-Stein identity and Square functions

An Overview of the Martingale Representation Theorem

THEOREMS, ETC., FOR MATH 515

Proofs of the martingale FCLT

Problem List MATH 5143 Fall, 2013

Functional Analysis Exercise Class

9 Brownian Motion: Construction

Fast-slow systems with chaotic noise

TEST CODE: PMB SYLLABUS

Chapter 8. P-adic numbers. 8.1 Absolute values

Learning Objectives for Math 165

Product measure and Fubini s theorem

How to Use Calculus Like a Physicist

Ito Formula for Stochastic Integrals w.r.t. Compensated Poisson Random Measures on Separable Banach Spaces. B. Rüdiger, G. Ziglio. no.

MATH MEASURE THEORY AND FOURIER ANALYSIS. Contents

The Pedestrian s Guide to Local Time

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

for all x,y [a,b]. The Lipschitz constant of f is the infimum of constants C with this property.

Homogenization for chaotic dynamical systems

Introduction to Topology

Stochastic integration. P.J.C. Spreij

1/12/05: sec 3.1 and my article: How good is the Lebesgue measure?, Math. Intelligencer 11(2) (1989),

4 Sums of Independent Random Variables

Notions such as convergent sequence and Cauchy sequence make sense for any metric space. Convergent Sequences are Cauchy

GAUSSIAN PROCESSES; KOLMOGOROV-CHENTSOV THEOREM

From Random Variables to Random Processes. From Random Variables to Random Processes

Stochastic Calculus. Alan Bain

Metric Spaces and Topology

Empirical Processes: General Weak Convergence Theory

THE INVERSE FUNCTION THEOREM

Dynkin (λ-) and π-systems; monotone classes of sets, and of functions with some examples of application (mainly of a probabilistic flavor)

Brownian motion. Samy Tindel. Purdue University. Probability Theory 2 - MA 539

NOTES ON DIOPHANTINE APPROXIMATION

Probability Theory I: Syllabus and Exercise

Weak convergence and Brownian Motion. (telegram style notes) P.J.C. Spreij

MA 8101 Stokastiske metoder i systemteori

Contents. 1 Preliminaries 3. Martingales

1. Stochastic Process

Summary of Real Analysis by Royden

Stochastic integral. Introduction. Ito integral. References. Appendices Stochastic Calculus I. Geneviève Gauthier.

SUMMARY OF RESULTS ON PATH SPACES AND CONVERGENCE IN DISTRIBUTION FOR STOCHASTIC PROCESSES

P (A G) dp G P (A G)

Math212a1413 The Lebesgue integral.

Generalized Gaussian Bridges of Prediction-Invertible Processes

BROWNIAN MOTION AND LIOUVILLE S THEOREM

MATH 6605: SUMMARY LECTURE NOTES

Normed and Banach spaces

Mathematical Methods for Physics and Engineering

FOUNDATIONS OF MARTINGALE THEORY AND STOCHASTIC CALCULUS FROM A FINANCE PERSPECTIVE

Lebesgue Integration: A non-rigorous introduction. What is wrong with Riemann integration?

Transcription:

ltcc5.tex Week 5 29 October 213 Ch. V. ITÔ (STOCHASTIC) CALCULUS. WEAK CONVERGENCE. 1. Quadratic Variation. A partition π n of [, t] is a finite set of points t ni such that = t n < t n1 <... < t n,k(n) = t; the mesh of the partition is π n := max i (t ni t n,(i 1) ), the maximal subinterval length. We consider nested sequences (π n ) of partitions (each refines its predecessors by adding further partition points), with π n. Call (writing t i for t ni for simplicity) π n B := t i π n (B(t i+1 ) B(t i )) 2 the quadratic variation of B on (π n ). The following classical result is due to Lévy (in his book of 1948); the proof below is from [Pro], I.3. THEOREM (Lévy). π n B t ( π n ) in mean square. Proof. π n B t = {(B(t i+i B(t i )) 2 (t i+1 t i )} = Y i, t i π n i {( i B) 2 ( i t)} = i say, where since i B N(, i t), E[( i B) 2 ] = t i, so the Y i have zero mean, and are independent by independent increments of B. So E[(π n B t) 2 ] = E[( i Y i ) 2 ] = i E(Y 2 i ), as variance adds over independent summands. Now as i B N(, i t), ( i B)/ i t N(, 1), so ( i B) 2 / i t Z 2, where Z N(, 1). So Y i = ( i B) 2 i t (Z 2 1) i t, E[(π n B t) 2 ] = i E[(Z 2 1) 2 ]( i t) 2 = c i ( i t) 2, writing c for E[(Z 2 1) 2 ], Z N(, 1), a finite constant. But since ( i t) 2 max i t. i t = π n.t, i i i E[(π n B t) 2 ] c.t. π n ( π n ). Note. 1. From convergence in mean square, one can always extract an a.s. convergent subsequence. 1

2. The conclusion above extends in full generality to a.s. convergence, but an easy proof needs the reversed mg convergence theorem (omitted). 3. There is an easy extension to a.s. convergence under the extra restriction n π n <, using the Borel-Cantelli lemma and Chebychev s inequality. 4. If we consider the theorem over [, t + dt], [, t] and subtract, we can write the result formally as (db t ) 2 = dt. This can be regarded either as a convenient piece of symbolism, or acronym, or as the essence of Itô calculus, to which we turn below. Background on Other Integrals. For simplicity, we fix T < and work throughout on time-set [, T ] (in financial applications, T is the expiry time, when or by when financial derivatives such as options to buy or sell expire. We want to define integrals of the form I t (f)(ω) = f(s, ω)db s, with suitable stochastic processes f as integrands and BM B as integrator. Remark. We first learn integration with x as integrator, to get f(x)dx, first as a Riemann integral (this is just the Sixth Form integral in the epsilon language of undergraduate mathematics), then as the Lebesgue integral (better, as more general, and easier to manipulate, thanks to the monotone and dominated convergence theorems, etc.). Later, e.g. in handling distribution functions, which may have jumps, we learn f(x)df (x) for F monotone. One can extend this by linearity to F a difference of two monotone functions a function locally of finite variation, FV. Again, such integrals fdf come in two kinds, Riemann-Stieltjes (R-S) and Lebesgue-Stieltjes (L-S). If we want fdf to exist for all continuous f as we do then one needs F to be FV (see e.g. [Pro], Th. 52 of I.7 though see [Mik], 2.1 for the surprising lengths to which one can push the R-S integral). Now BM has finite quadratic variation, so infinite ordinary variation. So one needs something new to handle BM as an integrator. 2. Itô Integral. We have our filtration (F t ) on Ω, where F t handles randomness up to time t, and the Borel σ-field B (the smallest containing the intervals) to handle the time-interval [, T ]. Write F t B for the smallest σ-field containing all A B, where A is F t -measurable and B is B-measurable. Call f(.,.) 2

measurable if it is F T B- measurable and adapted if f(t,.) is F t -measurable for each t [, T ]. Finally, write H 2 = H 2 [, T ] for the class of measurable adapted f satisfying the integrability condition E[ T f 2 (t, ω)dt] <. For (a, b] [, T ], f = I (a,b), the only candidate for I(f) := fdb is I(f)(ω) = b a db t = B b B a. Similarly for other kinds of interval, (a, b], [a, b], [a, b), since B t is continuous. Next, integration should be linear, so I should extend from indicators to simple functions by linearity. Write H 2 for the class of simple squareintegrable functions those f of the form n 1 f(t, ω) = a i (ω)i(t i < t t i+1 ) i= with a i F(t i )-measurable, E(a 2 i ) < and = t <... < t n = T. The only plausible candidate for I t (f) for f H 2 is n 1 I t (f)(ω) = a i (ω)(b(t t i+1 ) B(t t i )). i= By continuity of B t in t, I t (f) is continuous in t. Next, for u s, E(B u F s ) = B s as B is a mg, while for u s, E(B u F s ) = B u as then B u is known at time s. Combining, E(B u F s ) = B(min(u, s)). Thus for s t, n 1 E[I t (f)(ω) F s ] = a i (ω)(b(s t t i+1 ) B(s t t i ), which as s t is i= n 1 a i (ω)(b(s t i+1 ) B(s t i ), i= 3

which is I s (f). This says that I t (f) is a mg. So, from the mg property, the products of increments over disjoint intervals have zero mean (all relevant expectations exist as we are assuming square-integrability). Consider k 1 E[I t (f) 2 ] = E[( a i (B(t i+1 ) B(t i ))+a k (B(t) B(t k )) 2 ] (t k t < t k+1 ). i= Expand out the square; the cross-terms have zero expectation, leaving k 1 a 2 i (B(t i+1 ) B(t i )) 2 +a 2 k(b(t) B(t k )) 2 ] = E(a 2 i )(t i+1 t i )+E(a 2 k)(t t k ), k 1 E[ i= which is E(f 2 (s, ω))ds as f 2 is a 2 i on (t i, t i+1 ]. Combining: Proposition. For f H 2 a simple function, and (I t f)(ω) defined as above, (i) I t f is a continuous martingale, (ii) E[(I t f)(ω) 2 ] = E[ f 2 (s, ω)ds] (Itô isometry). One seeks to extend I t from simple functions f H 2 to general f H 2 (an extension analogous to the extension of the Lebesgue integral from simple to measurable functions, in the most basic non-random measure-theoretic setup). It is not at all obvious, but it is true, that with H 2, H 2 regarded as Hilbert spaces with the norm f := ( T f 2 (s, ω)ds) 1/2 H 2 is dense in H 2 - that is, each f H 2 is the limit in norm of an approximating sequence f n H 2. It is further true that the map I t extends from H 2 to H 2 via i= (I t f)(ω) := lim n (I t f n )(ω) (the limit is in the Hilbert-space norm), and that the limit above does not depend on the particular choice of approximating sequence. That is: Proposition. For f H 2 and (I t f)(ω) defined by approximation as above: (i) (I t f)(ω) is a continuous martingale, (ii) The Itô isometry E[(I t f)(ω) 2 ] = E[ f 2 (s, ω)ds holds. We must refer to a rigorous measure-theoretic treatment for details of the proof. Full accounts are in [R-Y], IV, [K-S], Ch. 3. See also [Pro], Ch. II. We now call I t f the Itô integral of f H 2, and use the suggestive Leibniz integral sign (which dates from 1675): (I t f)(ω) = f(s, ω)db(s, ω) 4 or f s db s ( t T ).

An alternative notation is (f.b) t := f s db s. Predictability. We used (t i, t i+1 ] to ensure left-continuity, so predictability. The distinction is not critical with Brownian motion or a continuous mg integrator, but it is critical with a general semimg integrator. The appropriate topology (or convergence concept, in metric-space situations as here) is that of uniform convergence in probability on compact time-sets (ucp). The simple predictable functions (class S) are dense in the adapted càglàd functions (class L) in the ucp topology ([Pro], II.4), and this allows the approximation and extension results we need (and assume). An Example. We calculate B sdb s, using Brownian motion both as (continuous, so previsible) integrand and as integrator. Take a sequence of partitions π n with mesh π n, and write t i for the partition points t ni of π n, as above. The approximation properties sketched above allow us to identify t B sdb s with the limit of B(t i )(B(t i+1 B(t i )). t i π n But this is 1 2 (B(t i+1)+b(t i ))(B(t i+1 ) B(t i )) 1 2 (B(t i+1) B(t i ))(B(t i+1 ) B(t i )). The first sum is 1 (B(ti+1 ) 2 B(t 2 i ) 2 ), which telescopes to 1 2 B(t)2 (B() = ). The second sum is 1 ( i B) 2, which tends to 1 t by Lévy s theorem on 2 2 the quadratic variation of BM. Combining: B s db s = 1 2 B2 t 1 2 t. This formula of course differs dramatically from that for ordinary (Newton- Leibniz) calculus, or its Riemann-Stieltjes or Lebesgue-Stieltjes extensions. The role of the second term the correction term or Itô term illustrates both the contrast between the Itô and earlier integrals and the decisive influence of the quadratic variation on the Itô integral. 5

Quadratic Variation. For continuous semimartingales X, the quadratic variation process X = ( X t : t ) is defined by X = X 2 2 X dx (of course X = X when X is continuous, as here). Alternatively, X 2 is a submg, and then X 2 = X + 2 X dx is the Doob (or Doob-Meyer) decomposition of X 2 into an increasing previsible process X and a mg 2 X dx. For general (not necessarily continuous) mgs X, the quadratic variation [X] involves the jumps X t of X also. Then if [X] c is the continuous part of the increasing process [X], [X] t = [X] c t + ( X s ) 2, s t and if X c is the continuous mg part of X in its semimg decomposition, X c = [X c ] = [X] c. Both X and [X] are shorthand for X, X and [X, X]. Extend to different X and Y by polarization, X, Y := 1 ( X + Y, X + Y X Y, X Y ) 4 (likewise for [.,.]); both are locally FV, so are semimgs ([Pro], II.6). Product Rule. The quadratic covariation [X, Y ] is, by polarization, [X, Y ] = XY X dy Y dx. Rearranging: for X, Y semimgs, so is XY, and then XY = X dy + Y dx + [X, Y ]. 6

This is the integration-by-parts formula, or product rule. It is the principal special case of Itô s formula (below) to which it is in fact equivalent. With H, K previsible integrands and X, Y semimg integrators, we can form both the stochastic integrals H.X = HdX, K.Y = KdY. We can then form the (square-)bracket processes, for which [H.X, K.Y ] t = H s K s d[x, Y ] s ; in particular, [H.X, H.X] t = Hs 2 d[x, X] s or [H.X] t = H 2 s d[x] s. For Brownian motion, [B] t = B t (as there are no jumps) = t (by Lévy s result on quadratic variation). So specialising, fdb = fs 2 ds, f s db s, g s db s = f s g s ds. 3. Itô s Formula The change-of-variable formula (or chain rule) of ordinary calculus extends to the Lebesgue-Stieltjes integral, and tells us that for smooth f ( C 1 ) and continuous A, of FV (on compacts), f(a t ) f(a ) = f (A s )da s. (This of course does not apply to B sdb s = 1 2 B2 t 1 t, but there the integral 2 is Itô, not Lebesgue-Stieltjes.) Rather less well-known is the extension to A only right-continuous: f(a t ) f(a ) = f (A s )da s + {f(a s ( f(a s ) f (A s ) A s }. + <s t One thus seeks a setting capable of handling both the last two displayed results together. Theorem (Itô s Formula). For X a semimartingale and f C 2, then f(x) is also a semimartingale, and f(x t ) f(x ) = f (X s )dx s + 1 f (X s )d[x, X] c s + 2 + + {f(x s ) f(x s ) f (X s ) X s }. <s t 7

First, we note the special case for X continuous. Theorem (Itô s Formula). For X a continuous semimartingale and f C 2, f(x) is also a continuous semimartingale, and f(x t ) f(x ) = f (X s )dx s + 1 2 f (X s )d[x, X] s. Corollary. If X = X + M + A is the decomposition of the continuous semimartingale X, that of f(x) is f(x t ) = f(x ) + f (X s )dm s + { f (X s )da s + 1 2 f (X s )d[m] s }. Proof. Let A be the class of C 2 functions f for which the desired result holds. Then A is a vector space. But A is also closed under multiplication, so is an algebra. Indeed, if f, g A, write F t, G t for the semimgs f(x t ) and g(x t ) and use the product rule. If X = X + M + A is the decomposition of X into a (continuous) local mg M and a BV process A, the continuous local mg part of F = f(x) is f (X s )dm s. So which by above is [F, G] t = [F cm, G cm ] t = [ The product rule now says that F t G t F G =. f (X s )dm s, f (X s )g (X s )d[m, M] s. F s dg s + G s df s + As Itô s formula holds for f (by assumption),. df t = f (X t )dx t + 1 2 f (X t )d[m] t, and similarly for g. Substituting, F t G t F G is {F s g (X s )+f (X s )G s }dx s + 1 2 g (X s )dm s ] t, (f g )(X s )d[m] s. {F s g (X s )+2f (X s )g (X s )+f (X s )G s }d[m] s. 8

This says that Itô s formula also holds for fg. So A is an algebra, and a vector space. Since A contains f(x) = x, A contains all polynomials. One can now reduce from the local-mg to the mg case by a localization argument, and extend from A to C 2 by an approximation argument. For details, see [R-W2], IV.32 (and for discontinuous X, VI.39). // Note. 1. Higher dimensions. For vector functions f = (F 1,..., f n ) and vector processes X = (X 1,..., X n ), we use the Einstein summation convention and D i f for f/ i. Then the Itô formula extends (with much the same proof) as f(x t ) f(x ) = D i f(x s )dx i s + 1 2 D ij f(x s )d[x i, X j ] s. 2. With f(x, y) = xy, we then recover the product rule (integration-by-parts formula). So Itô s formula and the product rule are equivalent. 3. So the class of semimgs is closed under C 2 functions: a powerful and highly non-linear closure property. 4. Differential Notation. In the one-dimensional case, we may re-write Itô s formula in shorthand form using differential rather than integral notation as df(x t ) = f (X t )dx t + 1 2 f (X t )d[x] t. The left is called the stochastic differential of f(x). So the above gives the stochastic differential of f(x) in terms of the stochastic differential dx of X and the quadratic variation [X, X] or [X] of X. 5. Formalism. For the continuous case, we can obtain Itô s formula from Taylor s formula in differential form, if we use these rules for differentials: dx i dx j = d[x i, X j ], dx i dx j dx k =, dxdv = whenever V has FV. In particular, for X = B BM in one dimension, db i tdb j t = (i j), (db i t) 2 = dt, dbdv = V of finite variation on compacts. The interpretation here is that for i j, the Brownian increments db i and db j are independent with zero mean, so E(dB i tdb j t ) = E(dB i t).e(db j t ) =. =, while for i = j the earlier symbolism (db t ) 2 = dt becomes (db i t) 2 = dt. The formalism above works well, and is flexible and reliable in practice. One can also do stochastic calculus for Lévy processes; see Applebaum [A]. Apart from BM, the prime case is the Poisson process; see Kingman [K]. 9

4. Weak convergence Convergence in Distribution. If c n, c are constants, and c n c, then regarding c n, c as random variables one should have c n c in any sense of convergence of random variables in particular, for convergence in distribution. If F cn, F c are their distribution functions, one has F cn (x) F c (x) as n for all x c the point where the limit function has a jump. For a set A, write Ā for its closure, Ao for its interior, A := Ā\Ao for its boundary. With P n, P the corresponding probability measures, P n ((, x]) P ((, x]) for all x / (, c]. It turns out that this is the right definition of convergence in distribution, as it generalizes to d dimensions (random vectors), and infinitely many dimensions (stochastic processes). We confine ourselves here to processes with continuous paths (e.g. Brownian motion). We take the (time) parameter set as [, 1] for simplicity. If X is such a process, its paths lie in C[, 1], the space of continuous functions on [, 1]. This is a metric space under the sup-norm metric d(f, g) := sup{ f(t) g(t) : t [, 1]}. The Borel σ-field B = B(C([, 1]) of C[, 1] is the σ-field generated by the open sets of C[, 1] w.r.t. this metric. The distribution, or law, of a process X on C[, 1] is given by P (B) := P (X(.) B) for B B. If X n, X are such processes, with laws P n, P, one says X n X in distribution, or in law, or weakly, or P n P weakly, if P n (B) P (B) (n ) B B with P ( B) =. It turns out that this is equivalent to fdp n fdp (n ) f C[, 1], and also, by Prohorov s theorem, to (i) convergence of finite-dimensional distributions (clearly needed), and (ii) tightness: for all ϵ > there exists a compact set K such that P n (K) > 1 ϵ for all n. For proof, see e.g. Billingsley [B]. Statistical Applications. These include the Kolmogorov-Smirnov (K-S) tests for equality of two distributions F, G, in terms of the Kolmogorov-Smirnov statistic D n := sup{ F n (.) G n (.)} of the distance between their empirical distribution functions. In fact D n has the same law as sup{b (t) : t [, 1]}, where B is Brownian motion and B is the Brownian bridge: B (t) := B(t) t. Similarly for many other functionals Donsker s Invariance Principle. This is the dynamic form of the Central Limit Theorem (CLT). 1