Stationary Gaussian Markov Processes As Limits of Stationary Autoregressive Time Series

Similar documents
Stationary Gaussian Markov processes as limits of stationary autoregressive time series

Differentiation in higher dimensions

HOMEWORK HELP 2 FOR MATH 151

Polynomial Interpolation

Financial Econometrics Prof. Massimo Guidolin

The derivative function

Lecture XVII. Abstract We introduce the concept of directional derivative of a scalar function and discuss its relation with the gradient operator.

1 The concept of limits (p.217 p.229, p.242 p.249, p.255 p.256) 1.1 Limits Consider the function determined by the formula 3. x since at this point

Mathematics 5 Worksheet 11 Geometry, Tangency, and the Derivative

Polynomial Interpolation

POLYNOMIAL AND SPLINE ESTIMATORS OF THE DISTRIBUTION FUNCTION WITH PRESCRIBED ACCURACY

Solutions to the Multivariable Calculus and Linear Algebra problems on the Comprehensive Examination of January 31, 2014

Volume 29, Issue 3. Existence of competitive equilibrium in economies with multi-member households

Analytic Functions. Differentiable Functions of a Complex Variable

A SHORT INTRODUCTION TO BANACH LATTICES AND

Symmetry Labeling of Molecular Energies

1. Introduction. Tis paper is concerned wit estimation of te parameters of a non-negative Levydriven Ornstein-Ulenbeck process and of te parameters of

Function Composition and Chain Rules

Exam 1 Review Solutions

NUMERICAL DIFFERENTIATION

WYSE Academic Challenge 2004 Sectional Mathematics Solution Set

1 Calculus. 1.1 Gradients and the Derivative. Q f(x+h) f(x)

University Mathematics 2

GELFAND S PROOF OF WIENER S THEOREM

On convexity of polynomial paths and generalized majorizations

Derivatives of Exponentials

MVT and Rolle s Theorem

Preface. Here are a couple of warnings to my students who may be here to get a copy of what happened on a day that you missed.

Quantum Numbers and Rules

LIMITS AND DERIVATIVES CONDITIONS FOR THE EXISTENCE OF A LIMIT

Math 161 (33) - Final exam

THE IDEA OF DIFFERENTIABILITY FOR FUNCTIONS OF SEVERAL VARIABLES Math 225

. If lim. x 2 x 1. f(x+h) f(x)

1.5 Functions and Their Rates of Change

3.4 Worksheet: Proof of the Chain Rule NAME

Combining functions: algebraic methods

Pre-Calculus Review Preemptive Strike

Math 31A Discussion Notes Week 4 October 20 and October 22, 2015

Continuity and Differentiability Worksheet

5.1 We will begin this section with the definition of a rational expression. We

158 Calculus and Structures

Generic maximum nullity of a graph

Solution. Solution. f (x) = (cos x)2 cos(2x) 2 sin(2x) 2 cos x ( sin x) (cos x) 4. f (π/4) = ( 2/2) ( 2/2) ( 2/2) ( 2/2) 4.

Definition of the Derivative

More on generalized inverses of partitioned matrices with Banachiewicz-Schur forms

4. The slope of the line 2x 7y = 8 is (a) 2/7 (b) 7/2 (c) 2 (d) 2/7 (e) None of these.

Cointegration in functional autoregressive processes

= 0 and states ''hence there is a stationary point'' All aspects of the proof dx must be correct (c)

Poisson Equation in Sobolev Spaces

Smoothness of solutions with respect to multi-strip integral boundary conditions for nth order ordinary differential equations

Lecture 21. Numerical differentiation. f ( x+h) f ( x) h h

How to Find the Derivative of a Function: Calculus 1

Test 2 Review. 1. Find the determinant of the matrix below using (a) cofactor expansion and (b) row reduction. A = 3 2 =

Quaternion Dynamics, Part 1 Functions, Derivatives, and Integrals. Gary D. Simpson. rev 01 Aug 08, 2016.

Strongly continuous semigroups

Semigroups of Operators

arxiv: v1 [math.pr] 28 Dec 2018

Functions of the Complex Variable z

lecture 26: Richardson extrapolation

Chapter 5 FINITE DIFFERENCE METHOD (FDM)

Click here to see an animation of the derivative

The Verlet Algorithm for Molecular Dynamics Simulations

ACCESS TO SCIENCE, ENGINEERING AND AGRICULTURE: MATHEMATICS 1 MATH00030 SEMESTER /2019

Differential Calculus (The basics) Prepared by Mr. C. Hull

NUMERICAL DIFFERENTIATION. James T. Smith San Francisco State University. In calculus classes, you compute derivatives algebraically: for example,

MATH1131/1141 Calculus Test S1 v8a

7.1 Using Antiderivatives to find Area

Consider a function f we ll specify which assumptions we need to make about it in a minute. Let us reformulate the integral. 1 f(x) dx.

DELFT UNIVERSITY OF TECHNOLOGY Faculty of Electrical Engineering, Mathematics and Computer Science

A = h w (1) Error Analysis Physics 141

SFU UBC UNBC Uvic Calculus Challenge Examination June 5, 2008, 12:00 15:00

232 Calculus and Structures

Chapter 2 Limits and Continuity

A Goodness-of-fit test for GARCH innovation density. Hira L. Koul 1 and Nao Mimoto Michigan State University. Abstract

New families of estimators and test statistics in log-linear models

SECTION 3.2: DERIVATIVE FUNCTIONS and DIFFERENTIABILITY

f a h f a h h lim lim

5.74 Introductory Quantum Mechanics II

, meant to remind us of the definition of f (x) as the limit of difference quotients: = lim

ERROR BOUNDS FOR THE METHODS OF GLIMM, GODUNOV AND LEVEQUE BRADLEY J. LUCIER*

Numerical Differentiation

An approximation method using approximate approximations

1watt=1W=1kg m 2 /s 3

Continuity and Differentiability of the Trigonometric Functions

SECTION 1.10: DIFFERENCE QUOTIENTS LEARNING OBJECTIVES

Continuous Stochastic Processes

Homework 1 Due: Wednesday, September 28, 2016

5 Ordinary Differential Equations: Finite Difference Methods for Boundary Problems

Fractional Derivatives as Binomial Limits

1. Questions (a) through (e) refer to the graph of the function f given below. (A) 0 (B) 1 (C) 2 (D) 4 (E) does not exist

THE IMPLICIT FUNCTION THEOREM

1 Lecture 13: The derivative as a function.

Practice Problem Solutions: Exam 1

Math Spring 2013 Solutions to Assignment # 3 Completion Date: Wednesday May 15, (1/z) 2 (1/z 1) 2 = lim

New Distribution Theory for the Estimation of Structural Break Point in Mean

Chapter 2. Limits and Continuity 16( ) 16( 9) = = 001. Section 2.1 Rates of Change and Limits (pp ) Quick Review 2.1

MATH1151 Calculus Test S1 v2a

Subdifferentials of convex functions

Bootstrap prediction intervals for Markov processes

Copyright c 2008 Kevin Long

Transcription:

Stationary Gaussian Markov Processes As Limits of Stationary Autoregressive Time Series Lawrence D. Brown, Pilip A. Ernst, Larry Sepp, and Robert Wolpert August 27, 2015 Abstract We consider te class, C p, of all zero mean stationary Gaussian processes, Y t, t (, ) wit p derivatives, for wic te vector valued process ( Y (0) t, Y (1) t ),..., Y (p) t, t 0 ( ) is a p + 1-vector Markov process, were Y (0) t = Y (t). We provide a rigorous description and treatment of tese stationary Gaussian processes as limits of stationary AR(p) time series. MSC 2010 Primary: 60G10, Secondary: 60G15 Keywords: Continuous autoregressive processes, stationary Gaussian Markovian processes, stocastic differential. equations 1 Introduction In many data-driven applications in bot te natural sciences and in finance, time series data are often discretized prior to analysis and are ten formulated using autoregressive models. Te teoretical and applied properties 1

of te convergence of discrete autoregressive ( AR ) processes to teir continuous analogs (continuous autoregressive or CAR processes) as been well studied by many matematicians, statisticians, and economists. See, for example, te works of [3, [4, [2, and [5. A special class of autoregressive processes are te discrete-time zero-mean stationary Gaussian Markovian processes on te line (, ). Te continuous time analogs of tese processes are documented in [9 (c.10) and [10 (pp. 207-212). For processes in tis class, te sample pats possess p 1 derivatives at eac value of t, and te evolution of te process following t depends only on te values of tese derivatives at t. Notationally, we term suc a process as a member of te class C p. For convenience, we will use te notation CAR(p)=C p. Te standard Ornstein-Ulenbeck processes is of course a member of C 1, and ence CAR(p) processes can be described as a generalization of te Ornstein-Ulenbeck process. It is well understood tat te Ornstein-Ulenbeck process is related to te usual Gaussian AR(1) process on a discrete-time index, and tat an Ornstein- Ulenbeck process can be described as a limit of appropriately cosen AR(1) processes (see [7). In an analogous fasion we sow tat processes in C p are related to AR(p) processes and can be described as limits of an appropriately cosen sequence of AR(p) processes. Section 2 begins by reviewing te literature on CAR(p) processes, recalling tree equivalent definitions of te processes in C p. Section 3 discusses ow to correctly approximate C p by discrete AR(p) processes. Tis construction is, to te best of our knowledge, novel. 2 Equivalent Descriptions of te Class C p We give ere tree distinct descriptions of processes comprising te class C p, wic are documented in [10 (p. 212), but in different notation. [10 prove (p.211-212) tat tese descriptions are equivalent ways of describing te same class of processes. Te first description matces te euristic description given in te introduction. Te remaining descriptions provide more explicit descriptions tat can be useful in construction and interpretation of tese processes. In all te descriptions Y = {Y (t)} symbolizes a zero-mean Gaussian process on t [0, ). 2

2.1 Tree Definitions (I) Y is stationary. Te sample pats are continuous and are p 1 times differentiable, a.e., at eac t [0, ) (Te derivatives at t = 0 are defined only from te rigt. At all oter values of t, te derivatives can be computed from eiter te left or te rigt, and bot rigt and left derivatives are equal). We denote te derivatives at t by Y (i) (t), i = 1,..., p 1. At any t 0 (0, ), te conditional evolution of te process given Y (t), t [0, t 0 depends only on te values of Y (i) (t 0 ), i = 0,..., p 1. Te above can be formalized as ( follows: let Y (0) t, Y (1) t,..., Y (p 1) t Itô vector diffusion process defined by te system of equations ), t 0 denote te values of a mean zero dy (i 1) t = Y (i) t dt, t > 0, i = 1, 2,..., p 1 p 1 (2.1) dy (p 1) t = a i+1 Y (i) t dt + σdw t i=0 for all t > 0, were σ > 0 and te coefficients {a j } satisfy conditions (2.3) and (2.4), below. Ten let Y (t) = Y (0) t. (II) Eac Y C p is given uniquely by a certain polynomial P (z) via te covariance of any suc Y as follows: E [Y s Y t = R(s, t) = r(t s) = e i(t s)z P (z) 2 dz. (2.2) P (z) is a complex polynomial of degree p + 1. It as positive leading coefficients and all complex roots ζ j = ρ j + iσ j, j = 0,..., p wit σ j > 0, and ρ j real. We impose te following constraint on te roots of P (z): wenever ρ j 0, ten tere is anoter ζ j = ζ j wic is te negative conjugate of ζ j. Tis definition of P (z) ensures tat P (z) 2 is an even function. Finally, it can easily be sown tat, for all t, r(t) automatically as 2p derivatives. Te conditions in equations (2.3) and (2.4) below link equations (2.2) and (2.1) and caracterize wic processes satisfying (2.1) are stationary. Te coefficients a i are te unique solution of te equations: 3

r (p+i+1) (0 + ) = a j r (i+j) (0), i = 0, 1,..., p 1. (2.3) j=0 Note tat te left and rigt derivatives of r (j) are equal except for j = 2p. Te diffusion coefficient, σ, is given by σ 2 = a j r (j+p) (0)( 1) j+1 + ( 1) p r (2p+1) (0 ). (2.4) j=0 Te stationarity of te process in (2.1) can also be determined via te caracterization in Section 2.2. (III) Equivalently, it is necessary and sufficient tat Y C p as te representation via Wiener integrals wit a standard Brownian motion, W, Y t = g(t u)dw (u), (2.5) were g as te L 2 Fourier transform ĝ(z) = 1 P (z). (2.6) By well-known results from bot Fourier analysis and stocastic integration, a full treatment of wic is given in [6, (2.5) and (2.6) jointly are equivalent to construction (II). Furter, by [6, an equivalent construction to tat given jointly by equations (2.5) and (2.6) is te following: given a pair of independent standard Brownian motions, W 1 and W 2, Y as te following spectral representation: Y t = cos tzĝ dw 1 (z) + sin tzĝ dw 2 (z). (2.7) Wit regard to initial conditions for (2.1), we note tat te process in (2.2) as a stationary distribution. If we use tat distribution as te initial distribution for (2.1), and ceck tat equations (2.3) and (2.4) old, we arrive at a stationary process. 4

2.2 Caracterization of Stationarity via (2.1) Te system in (2.1) is linear. Stationary of vector-valued processes described in suc a way as been studied elsewere. See in particular ([7, p. 357, Teorem 5.6.7). Te coefficients in (2.1) tat yield stationarity can be caracterized via te caracteristic polynomial of te matrix Λ, were Λ λi is: λ p a p λ p 1... a 2 λ a 1 = 0. (2.8) Te process is stationary if and only if all te roots of equation (2.8) ave strictly negative real parts. In order to discover weter te coefficients in (2.1) yield a stationary process it is tus necessary and sufficient to ceck weter all te roots of equation (2.8) ave strictly negative real parts. In te case of C 2 te condition for stationarity is quite simple, namely tat a 1, a 2 sould lie in te quadrant a i < 0, i = 1, 2. Te covariance functions for C 2 can be found in [9 (p. 326). For iger order order processes te conditions for stationarity are not so easily described. Indeed, for C 3 it is necessary tat a i < 0, i = 1, 2, 3, but te set of values for wic stationarity olds is not te entire octant. For larger p one needs to study te solutions of te iger order polynomial in equation (2.8). 3 Weak Convergence of te -AR(2) Process to CAR(2) Process 3.1 Discrete Time Analogs of te CAR Processes We now turn our focus to describing te discrete time analogs of te CAR processes and te expression of te CAR processes as limits of tese discrete time processes. In tis section, we discuss te situation for p = 2. Define te -AR(2) processes on te discrete time domain domain {0,, 2,...} via X t = b 1X t + b 2X t 2 + ς Z t, (3.1) Z t IID N(0, 1), t = 2, 3,... Te goal is to establis conditions on te coefficients b 1, b 2 and ς so tat tese AR(2) processes converge to te continuous time CAR(2) process as in 5

te system of equations given in (2.1). We ten discuss some furter features of tese processes. To see te similarity of te -AR(2) process in (3.1) wit te CAR(2) process of (2.1), we introduce te corresponding -VAR(2) processes { 0;t, 1;t} defined via 0;t 0;t = 1;t, From (3.2) we see tat 1;t 1;t = [ c 1 0;t + c 2 1;t + ξ Z t. Z t IID N(0, 1), t =, 2,... (3.2) 0;t = 0;t + 1;t = 0;t + 1;t + [ c 1 0;t + c 2 1;t 2 + ξ Z t = [ 2 + c 1 2 + c 2 0;t [ 1 + c 2 0;t 2 + ξ Z t. Tis sows tat te -AR(2) process of (3.1) is equivalent to te -VAR(2) in (3.2) wit b 1 c 1 2 + c 2 + 2, b 2 c 2 1, and ς ξ, (3.3) or, equivalently, c 1 2 [ b 1 + b 2 1, c 2 1 [ 1 b 2, and ξ 1 ς. (3.4) From above, te -AR(2) process of (3.1) wit coefficients given by (3.3) is equivalent to te -VAR(2) in equations (3.2). Teorem 3.1. Consider a sequence of -AR(2) processes of (3.1) wit coefficients given by (3.3), were c j a j, j = 1, 2, and ξ / σ as 0. Tis sequence converges in distribution to te CAR(2) process of (2.1). Proof. It suffices to sow tat te -VAR(2) process of (3.2) converges to te SDE system of dy t = Ẏtdt, t > 0 dẏt = [ a 1 Y t + a 2 Ẏ t dt + σdwt, t > 0. (3.5) 6

We employ te framework of Teorems 2.1 and 2.2 of [8. Let M t be te σ-algebra generated by and 0;0, 0;, 0;2,..., 0;t 1;0, 1;, 1;2,..., 1;t for t =, 2,.... Te -VAR(2) process of (3.2) is clearly Markovian of order 1, since to construct { 0;t, 1;t} from { 0;t, 1;t } one needs to use te second equation of (3.2) to construct 1;t and ten use te first equation to construct 0;t as well. Tis establises tat 0;t is M t adapted. Tus, te corresponding drifts per unit of time conditioned on information at time t are given by: [ 0;t [ 0;t E M 0;t + 1;t 0;t t = E M t = 1;t (3.6) and [ 1;t+ [( ) 1;t c E M 1 0;t + c 2 1;t + ξ Z t+ t = E M t = c 1 0;t + c 2 1;t. (3.7) Furtermore, te variances and covariances per unit of time are given by [( 0;t 2 [( ) 0;t ) 2 E M 1;t t = E M t = ( 1;t) 2 (3.8) and [ ( 1;t+ 1;t) 2 [ (c ) 2 E M 1 0;t + c 2 1;t + ξ Z t+ t = E M t = [ ( c 1 0;t + c 21;t 2 [ + 2 c 1 0;t + c 21;t ) ξ 2 ξ E [Z t+ + E [ Zt+ 2 = [ ( ) c 1 0;t + c 2 2 ξ 2 1;t +, (3.9) 7

were te last equality assumes tat ɛ t+ IID N(0, 1). By te same logic, [( )( 0;t 0;t 1;t+ 1;t) E M t [ (c ) = E 1;t 1 0;t + c 2 1;t + ξ Z t+ M t (3.10) [ = 1;t c 1 0;t + c 21;t. Te relationsips in (3.8) - (3.10) become [( 0;t 2 0;t ) E M t = o(1), (3.11) [( 1;t+ 2 1;t) E M t = ( ξ ) 2 + o(1), (3.12) and [( )( 0;t 0;t 1;t+ 1;t) E M t = o(1). (3.13) Te o(1) terms vanis uniformly on compact sets. We may additionally sow by brute force tat te limits of [( 0;t 4 0;t ) E M t and [( 1;t+ 4 1;t) E M t exist and converge to zero as 0. We proceed to define te continuous time version of te -VAR(2) process of (3.2) by 8

0;t 0;k and 1;t 1;k for k t < (k + 1). Ten, according to te Teorem 2.2 in [8, te relationsips (3.6), (3.7) and (3.11) - (3.13) provide te weak (in distribution) limit diffusion [ [ [ [ [ Yt 0 1 Yt 0 0 0 d = dt + d, a 1 a 2 0 σ W t Ẏ t Ẏ t were W t, t 0, is a Brownian motion. Tis is te linear SDE system of (3.5) and it as a unique solution. Remark 3.1. [8 Teorems 2.1 and 2.2 are explicitly stated for real-valued processes but apply to vector valued processes as well. One only needs to explicitly allow te processes to be vector valued, and to write te regularity conditions to allow for te full cross-covariance of te vector valued observations, rater tan just ordinary covariance functions. Our processes are muc better beaved tan te most general type of process [8 considers since our error variance is constant (depending only on ) and our distributions are Gaussian, and ence very ligt tailed. Tus bot of Teorems 2.1 and 2.2 in [8 apply. 3.2 Stationarity of AR(2)-VAR(2) process In tis section we present necessary and sufficient conditions for stationarity of te AR(2) process and its equivalent VAR(2) process and connect tis to te stationary condition for CAR(2). First, we invoke te following proposition, wic is easily proved using well-known results from te time series literature (see [1). Proposition 3.1. Te AR(2) process X t = b 1 X t 1 + b 2 X t 2 + ς Z t, Z t IID N(0, 1). (3.14) is stationary if and only if 2 < b 1 < 2, b 2 b 1 < 1 and b 1 + b 2 < 1, (3.15) wic require ( ) b 1, b 2 to lie in te interior of te triangle wit vertices (-2,-1), (0,1) and (2,-1). Its stationary variance is given by γ 0 = ς 2 1 b2 1 1 b 2 b 2 ( b 2 1 1 b 2 + b 2 ). (3.16) 9

Because of te relations in (3.3) and (3.4), it follows tat we ave te following corollary to Proposition 3.1. Corollary 3.1. Te VAR(2) process 0;t 0;t 1 = 1;t, 1;t 1;t 1 = c 1 0;t 1 + c 2 1;t 1 + ξz t, Z t IID N(0, 1), t = 1, 2,..., is equivalent to te AR(2) process of (3.14) for b 1 = c 1 + c 2 + 2 b 2 = c 2 1 and ς = ξ. Furtermore, te VAR(2) process is stationary if and only if 4 < c 1 < 0 and 2 c 1 2 < c 2 < 0. (3.17) (3.17) is equivalent to te condition tat (c 1, c 2 ) lies in te interior of te triangle wit vertices (-4,0), (0,0) and (0,-2). 3.3 Stationary Variance of -VAR(2) Process In tis section we investigate te stationarity of a special version of te - VAR(2) process in Teorem 5.1.1, wic converges to te CAR(2) process of (3.5) as 0, troug te stationarity of its equivalent -AR(2) process. Proposition 3.2. Te -VAR(2) process of (3.2) for c 1 = a 1, c 2 = a 2 and ξ = σ is stationary as 0 if and only if a j < 0, j = 1, 2. Ten its stationary variance satisfies lim γ0 = 0 σ2 2a 1 a 2. (3.18) Proof. From (3.3), te -VAR(2) process of te ypotesis is equivalent to te -AR(2) process of equation (3.1) wit coefficients given by b 1 = a 1 2 + a 2 + 2, b 2 = a 2 1 and ς = σ ( ) 3. (3.19) From condition (3.15) of Proposition 3.1, tis -AR(2) process is stationary if and only if te following conditions old: b 1 + b 2 < 1 a 1 2 + a 2 + 2 a 2 1 < 1 a 1 < 0, 10

b 1 < 2 a 1 2 + a 2 + 2 < 2 a 2 < a 1 0 a 2 < 0, 2 < b 1 2 < a 1 2 + a 2 + 2 a 1 2 + a 2 + 4 > 0 0 < < a 2 a 2 2 16a 1 2a 1, b 2 b 1 < 1 a 2 1 a 1 2 a 2 2 < 1 a 1 2 + 2a 2 + 4 > 0 0 < < a 2 a 2 2 4a 1 a 1, were te last two conditions old as 0. Finally, from equation (3.16) and (3.19) we can compute te stationary variance as follows: ( ) ς 2 γ0 = ( ) 2 [( ) 2, b 1 1 b b 1 2 + b 2 1 b 2 1 b 2 = σ 2 (2 + a 2 ) a 1 a 2 (4 + a 1 2 + 2a 2 ) 0 = σ2. 2a 1 a 2 Tis concludes te proof. 3.4 Stationarity of CAR(2) Process Tis section provides a new derivation of te necessary and sufficient conditions for te stationarity of a CAR(2) ( process. ) It also gives te stationary covariance function of te vector Y (0) t, Y (1) t. Teorem 3.2. Te CAR(2) process given by te SDEs system [ [ [ Yt Yt 0 d = Λ dt + Σ d, t > 0, (3.20) Ẏ t Ẏ t W t were [ 0 1 Λ = a 1 a 2 and Σ = [ 0 0 0 σ, 11

is stationary if and only if a j < 0, j = 1, 2. Ten its solution [Y t, Ẏt, t 0, is a zero-mean 2-dimensional Gaussian process wit covariance V 0 e tλ ΣΣ e tλ dt (3.21) and covariance function ([ Ys [ Ẏt ) ρ(s, t) E Y t, Ẏs { e (s t)λ V, 0 t s < ; = V e (t s)λ, 0 s t <. Proof. According to Teorem 6.7 in Capter 5 of [7, te assertion of te teorem olds if all te eigenvalues of matrix Λ ave negative real parts. Hence, we compute te eigenvalues of matrix Λ. We calculate te caracteristic polynomial as λ 1 φ(λ) = Λ λi = a 2 λ = λ2 a 2 λ a 1, a 1 wic is of quadratic order wit a discriminant equal to D = a 2 2 + 4a 1. We ten consider te following cases: (i) If D = 0 a2 2 4 = a 1, te caracteristic polynomial as te double root λ 1,2 = a 2 2. (ii) If D > 0 a2 2 4 > a 1, te caracteristic polynomial as te two real roots λ 1,2 = a 2 ± a 2 2 + 4a 1. 2 (iii) If D < 0 a2 2 4 < a 1, te caracteristic polynomial as te two complex roots λ 1,2 = a 2 ± i 4a 1 + a 2 2. 2 In every case we need to impose conditions on te coefficients of te caracteristic polynomial so as te real part of all eigenvalues is negative. 12

Indeed, in case (i) te double real root of te caracteristic polynomial is negative if and only if a 2 < 0, wic troug te discriminant condition implies also tat a 1 < 0. In case (ii) we need to impose tat bot real eigenvalues are negative; i.e., λ 1 = a 2 a 2 2 + 4a 1 2 < 0, λ 2 = a 2 + a 2 2 + 4a 1 < 0 2 a 2 2 + 4a 1 < a 2 a 2 <0 a 2 2 + 4a 1 < a 2 2 a 1 < 0. Note tat te latter condition implies bot tat a j < 0 for j = 1, 2. Ten te former condition olds as well. In case (iii) te common real part of te two complex eigenvalues is negative if and only if a 2 < 0, wic troug te discriminant condition also implies tat a 1 < 0. Consequently, in all cases te real part of bot eigenvalues of matrix Λ is negative if and only if a j < 0, j = 1, 2. We now compute te stationary variance V of te CAR(2) process of (3.20), as given in (3.21), beginning wit te computation of te matrix e tλ, t 0. In particular, we are looking for f(λ), were f(λ) = e λt. From standard matrix teory, tis can be computed via a polynomial expression of order 1, and tus f(λ) = δ 0 I + δ 1 Λ. Hence, it suffices to set g(λ) = δ 0 + δ 1 λ and to demand tat f(λ) and g(λ) to be equal on te spectrum of Λ. Ten we will ave tat f(λ) = g(λ). Te roots (one double or two real/complex) λ 1, λ 2 of te caracteristic polynomial φ(λ) of Λ satisfy te relationsips: λ 1 + λ 2 = a 2 and λ 1 λ 2 = a 1. (3.22) Since te polynomials f(λ) = e λt and g(λ) = δ 0 + δ 1 λ must be equal on te spectrum of Λ, we ave tat f(λ 1 ) = g(λ 1 ) e λ 1 t = δ 0 + δ 1 λ 1, f(λ 2 ) = g(λ 2 ) e λ 2 t = δ 0 + δ 1 λ 2. 13

Ten, e Λt = f(λ) = g(λ) = δ 0 I + δ 1 Λ = δ 0 δ 1. a 1 δ 1 δ 0 + a 2 δ 1 From (3.21), we compute te stationary variance V = 0 e tλ ΣΣ e tλ dt = σ 2 0 e tλ [ 0 0 0 1 (e tλ ) dt. For any a < 0 we ave tat e (a+bi)t dt = e(a+bi)t (a + bi) 0 0 = 1 a + bi. Using te equalities in (3.22), we can rewrite te covariance function as: = σ 2 λ 2e λ 1 (s t) λ 1 e λ 2 (s t) 2a 1 a 2 (λ 2 λ 1 ) σ 2 eλ 2 (s t) e λ 1 (s t) 2a 2 (λ 2 λ 1 ) σ 2 eλ 2 (s t) e λ 1 (s t) 2a 2 (λ 2 λ 1 ) σ 2 λ 2e λ 2 (s t) λ 1 e λ 1 (s t) 2a 2 (λ 2 λ 1 ), 0 t s <. 3.5 Weak Convergence of -AR(p) process to CAR(p) process We now consider te AR(p) process on te discrete time domain {0,, 2,...}, given as X t = b 1X t + b 2X t 2 + + b i X t i + + b px t p + ς Z t, (3.23) Z t IID N(0, 1), t = p, (p + 1),..., and sow tat subject to suitable conditions on te coefficients b 1, b 2,... b p and ς, tis converges as 0 to its continuous time CAR(p) process of te form p 1 Y (p) t = a i+1 Y (i) t + σw t, t > 0, (3.24) i=0 14

for a j 0, j = 1, 2,..., p, and σ 2 > 0. Define te coefficients {c j : j = 1,..., p} and ζ troug te equations { (p ) ( ) } k 1 b i ( 1) i 1 + p k+1 c k, and (3.25) i i 1 ς p 1 ξ. k=i Te following teorem, wic is relegated to te Appendix, can be proven: Teorem 3.3. Te -AR(p) process of (3.23) wit coefficients given by (3.25), were c j a j, j = 1, 2,..., p, and ξ / σ as 0, converges in distribution to te CAR(p) process of (3.24). It is of interest to note te scaling for te Gaussian variable Z t in (3.23). In order to ave te desired convergence, one must ave ζ σ and via (3.25 tis entails ζ σ p 1/2. APPENDIX Te Appendix proves Teorem 5.3. To do so, we first study te similarity of te -AR(p) process in (3.23) wit te CAR(p) process (3.24). We begin by introducing te corresponding -VAR(p) process 0;t 0;t = 1;t, 1;t 1;t = 2;t, i 1;t i 1;t = i;t,. (A.1) p 2;t p 2;t = p 1;t,. p 1 p 1;t p 1;t = c i+1 i;t + ξ Z t, i=0 Te process of (A.1) immediately yields: i=1 = 1;t = 0;t 0;t 2 Z t IID N(0, 1), t =, 2,... 15

= ( ) 1 ( 1) 0 0;t + 0 i=2 = 2 2;t = [ 1;t 1;t 2 ( ) 1 ( 1) 1 1 0;t 2, = [ 0;t 0;t 2 [ 0;t 2 1;t 3 = 0;t 2 0;t 2 + 0;t 3 ( ) ( ) ( ) 2 2 2 = ( 1) 0 0;t + ( 1) 1 0;t 2 + ( 1) 2 0 1 2 0;t 3. Te process of (A.1) generalizes as follows: i+1 ( ) i i i;t = ( 1) k 1 k 1 0;t k, k=1 (A.2) for i = 1, 2,..., p 1. We prove (A.2) via matematical induction. (i) For i = 1 te relationsip (A.2) olds trivially. (ii) Let (A.2) old for i = m. (iii) We sall sow tat (A.2) also olds for i = m + 1. m+1 m+1;t (A.1) for i=m+1 =========== m m;t m m;t 2 m+1 (ii) === k=1 m+1 k=1 ( ) m ( 1) k 1 k 1 0;t k ( ) m ( 1) k 1 0;t (k+1) k 1 m+1 2nd sum: l=k+1 =========== = m+2 k=1 k=1 m+2 l=2 ( ) m ( 1) k 1 k 1 0;t k+ ( ) m ( 1) l 1 0;t l l 2 ( ) m + 1 ( 1) k 1 k 1 0;t k. 16

Furtermore, we ave tat 0;t (A.1) for i=1 ========= 0;t + 1;t (A.1) for i=2 ========= 0;t + [ 1;t + 2;t = 0;t + 1;t + 2 2;t p 2 = = i i;t + p 1 p 1;t i=0 p 2 by te last equation of (A.1) ================== p 1 (A.2) ==== i=0 i i;t + [ p 1 p 1;t + i=0 p 1 c i+1 i;t + ξ Z t i=0 [ i+1 1 + p i c i+1 by intercanging te sums =============== k=1 ( ) i ( 1) k 1 0;t k + p 1 ξ Z t k 1 { p 1 ( 1) k 1 k=1 i=k 1 0;t k + p 1 ξ Z t ( ) i [ } 1 + p i c i+1 k 1 telescopic sum ========== { (p ) ( 1) k 1 + k k=1 p 1 ξ Z t, i=k ( ) } i 1 p i+1 c i k 1 0;t k+ wic, wen compared wit te -AR(p) process of (3.23), yields te relationsips { (p ) ( ) } k 1 b i ( 1) i 1 + p k+1 c k, and (A.3) i i 1 ς p 1 ξ k=i 17

for all i = 1, 2..., p. From above, te -AR(p) process of (3.23) wit coefficients given by (3.25) is equivalent to te -VAR(p) of (A.1). We conclude wit a statement of Teorem 5.3, te proof and supporting details of wic are in te Appendix. We sall next find te coefficients c i, i = 1, 2,..., p, in terms of te coefficients b i, i = 1, 2,..., p. In particular, we ave tat, from (3.25), {( ) ( ) } i=p p p 1 = b p = ( 1) p 1 + c p p p 1 [ ( ) p 1 c p = 1 ( 1) p 1 b p 1 p 1 and { ( ) i=p 1 p = b p 1 = ( 1) p 2 + p 1 k=p 1 { [( ) p 2 c p 1 = 2 ( 1) p 2 b p 1 + p 2 ( ) } k 1 p k+1 c k p 2 ( ) } p 1 b p 1. p 2 Proposition A.1. Tis gives te general formula [ ( ) k 1 c i = p+i 1 ( 1) i 1 b k 1, i = 1, 2,..., p. (A.4) i 1 k=i Proof. We prove Proposition A.1 by backward induction. (i) For i = p te relationsip (A.4) olds trivially. (ii) Let (A.4) old for every i = p 1, p 2,..., m + 1. (iii) We sall sow tat (A.4) also olds for i = m. { ( ) i=m p = b m = ( 1) m 1 + m i=m+1 i=m ( ) (i), (ii) p = ( 1) m 1 b m = + p m+1 c m m+ ( ) i 1 ( 1) i 1 m 1 ( ) i 1 p i+1 c i m 1 k=i } ( ) k 1 b k i 1 i=m+1 ( ) i 1. m 1 Hence, intercanging te order of summation in te double sum, te former 18

index bounds i k p and m+1 i p ave now become m+1 i k and m + 1 k p. We furter ave: as well as i=m+1 ( ) i 1 ( ) k 1 ( 1) i 1 b k = m 1 i 1 i=m+1 k=i k ( )( ) i 1 k 1 = b k ( 1) i 1 m 1 i 1 k=m+1 j=i m ====== i=m+1 k=m+1 b k = ( 1) m 1 k=m+1 ( ) k m k 1 ( ) k m ( 1) j+m 1 m 1 j b k Newton ====== ( 1) m 1 = ( 1) m ( ) i 1 = m 1 k=m+1 i=m+1 Te last relationsip yields b k j=1 ( ) [ k m k 1 ( ) k m ( 1) j m 1 j k=m+1 b k ( ) k 1, m 1 [( ) i m j=0 1 ( ) k 1 [ ( 1 + 1) k m 1 m 1 ( ) i 1 m telescopic sum ========== ( ) p ( 1) m 1 b m = + p m+1 c m + ( 1) m m [( ) p 1 m ( ) k 1 p m+1 c m = ( 1) m 1 b k 1, m 1 k=m concluding part (iii) and tus te proof. 19 k=m+1 b k ( ) p 1. m ( ) k 1 m 1

Finally, for t > 0 we know tat: dy t = Y (1) t dt, dy (1) dy (i 1) t dy (p 2) t t = Y (2) t dt,.. = Y (i) t dt, (A.5) = Y (p 1) t dt, and from (3.24) we ave also tat [ dy (p 1) t = a 1 Y t + a 2 Y (1) t + + a i Y (i 1) t + + a p 1 Y (p 2) t + a p Y (p 1) t dt + σdw t. (A.6) Tus te CAR(p) process of (3.24) is equivalent from (A.5) and (A.6) to te system of stocastic differential equations in (2.1). Teorem 5.3 Te -AR(p) process of (3.23) wit coefficients given by (3.25), were c j a j, j = 1, 2,..., p, and ξ / σ as 0, converges in distribution to te CAR(p) process of (3.24). Proof. We generalize Teorem 3.3 by proving tat it suffices to sow tat te -VAR(p) process of (A.1) converges to te SDEs system of (2.1). As in Teorem 5.1.1, we employ te framework of Teorems 2.1 and 2.2 of [8. Let M t be te σ-algebra generated by i;0, i;, i;2,..., i;t, i = 0, 1,..., p 2, and p 1;0, p 1;, p 1;2,..., p 1;t for t =, 2,... Te -VAR(p) process of (A.1) is clearly Markovian of order 1, since we may construct 0;t, 1;t, 2;t,..., p 1;t from 0;t, 1;t, 2;t,..., p 1;t by constructing first p 1;t from te last equation of (A.1), p 2;t from (A.1) for i = p 1, and so fort, and ten finally 0;t from (A.1) for i = 1. Tis also establises tat i;t, i = 0, 1,..., p 2 is M t adapted. Tus te corresponding drifts per unit of time conditioned on information at time t are given by: [ i 1;t [ i 1;t E M (A.1) i 1;t + i;t i 1;t t ==== E M t 20

= i;t, i = 1, 2,..., p 1, (A.7) and [ p 1;t+ p 1;t E M t [ by te last equation of (A.1) p 1 i=0 ================== E c i+1 i;t + ξ Z t+ M t = c 1 0;t + c 2 1;t + + c p p 1;t. (A.8) Furtermore, te variances and covariances per unit of time are given by ( ) 2 ( ) i 1;t 2 i 1;t E M t ==== (A.1) i;t E M t ( 2, = i;t) i = 1, 2,..., p 1, (A.9) and ( p 1;t+ p 1;t E ) 2 M t [ by te last equation of (A.1) 2 p 1 i=0 c i+1 i;t) + ξ Z t+ ================== E M t = [ c 1 0;t + c 2 1;t + + c p p 1;t 2 + ( ξ ) 2, (A.10) were te last equality assumes tat Z t+ IID N(0, 1). By te same logic: ( )( ) i 1;t i 1;t j 1;t j 1;t E M t 21

and ( )( ) (A.1) i;t j;t ==== E M t = i;t j;t, i, j = 1, 2,..., p 1, i j, (A.11) ( )( ) i 1;t i 1;t p 1;t+ E p 1;t M t [ (c ) (A.1) ==== E i;t 1 0;t + c 2 1;t + + c p p 1;t + ξ Z t+ M t [ = i;t c 1 0;t + c 2 1;t + + c p p 1;t, i = 1, 2,..., p 1. (A.12) Terefore, te relationsips of (A.9) - (A.12) become ( i 1;t i 1;t E ) 2 M t = o(1), i = 1, 2,..., p 1, (A.13) ( p 1;t+ p 1;t E ) 2 M t = ( ξ ) 2 + o(1), (A.14) ( )( E i 1;t i 1;t j 1;t j 1;t ) M t = o(1) (A.15) and ( )( E i, j = 1, 2,..., p 1, i j i 1;t i 1;t p 1;t+ p 1;t ) M t = o(1), (A.16) 22

i = 1, 2,..., p 1, were te o(1) terms vanis uniformly on compact sets. We may also sow by brute force tat te limits of ( ) 4 i 1;t i 1;t E M t, i = 1, 2,..., p 1 and ( p 1;t+ p 1;t E ) 4 M t exist and converge to zero as 0. We can ten define te continuous time version of te -VAR(p) process of (A.1) by i;t i;k for k t < (k + 1) and i = 0, 1,..., p 1. Tus, according to Teorem 2.2 in [8, te relationsips (A.7), (A.8) and (A.13) - (A.16) provide te weak (in distribution) limit diffusion. Tis is precisely te linear SDE system of (2.1) and it as a unique solution. References [1 Anderson, T. Te Statistical Analysis of Time Series. Wiley- Interscience, 1994. [2 Brockwell, P. J., Ferrazzano, V., and Klüppelberg, C. Higfrequency sampling and kernel estimation for continuous-time moving average processes. J. Time Series Anal. 34, 3 (2013), 385 404. [3 Brockwell, P. J., and Lindner, A. Existence and uniqueness of stationary Lévy-driven CARMA processes. Stocastic Process. Appl. 119, 8 (2009), 2660 2681. [4 Brockwell, P. J., and Lindner, A. Strictly stationary solutions of autoregressive moving average equations. Biometrika 97, 3 (2010), 765 772. 23

[5 Brockwell, P. Davis, R., and Yang, Y. Continuous-time Gaussian autoregression. Statistica Sinica 17, 1 (2007), 63. [6 Dym, H., and McKean, H. Gaussian Processes, Function Teory, an te Inverse Spectral Problem. Academic Press, 1976. [7 Karatzas, I., and Sreve, S. Brownian Motion and Stocastic Calculus. Springer, 1991. [8 Nelson, D. ARCH models as diffusion approximations. Journal of Econometrics 45, 1 (1990), 7 38. [9 Papoulis, A. Probability, Random Variables, Stocastic Processes. Mc- Graw Hill, 1991. [10 Rasmussen, C., and Williams, C. Te Statistical Analysis of Time Series. MIT Press, 2006. 24