Continuous-time Gaussian Autoregression Peter Brockwell, Richard Davis and Yu Yang, Statistics Department, Colorado State University, Fort Collins, CO

Similar documents
Numerical methods for solving stochastic differential equations

The Cameron-Martin-Girsanov (CMG) Theorem

The multidimensional Ito Integral and the multidimensional Ito Formula. Eric Mu ller June 1, 2015 Seminar on Stochastic Geometry and its applications

Introduction to Diffusion Processes.

Gaussian processes. Basic Properties VAG002-

Ulam Quaterly { Volume 2, Number 4, with Convergence Rate for Bilinear. Omar Zane. University of Kansas. Department of Mathematics

March 16, Abstract. We study the problem of portfolio optimization under the \drawdown constraint" that the

Asymptotic inference for a nonstationary double ar(1) model

Strictly Stationary Solutions of Autoregressive Moving Average Equations

Monte Carlo Methods for Statistical Inference: Variance Reduction Techniques

ON THE FIRST TIME THAT AN ITO PROCESS HITS A BARRIER

Bernardo D Auria Stochastic Processes /12. Notes. March 29 th, 2012

1. Introduction. Tis paper is concerned wit estimation of te parameters of a non-negative Levydriven Ornstein-Ulenbeck process and of te parameters of

An Analytic Method for Solving Uncertain Differential Equations

Mean-square Stability Analysis of an Extended Euler-Maruyama Method for a System of Stochastic Differential Equations

Stochastic Differential Equations

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

A Note on the Non-negativity of Continuous-time ARMA and GARCH Processes

1 Brownian Local Time

A NOTE ON STOCHASTIC INTEGRALS AS L 2 -CURVES

REVIEW OF DIFFERENTIAL CALCULUS

The concentration of a drug in blood. Exponential decay. Different realizations. Exponential decay with noise. dc(t) dt.

Supplement A: Construction and properties of a reected diusion

Bootstrapping continuous-time autoregressive processes

Time Series Analysis -- An Introduction -- AMS 586

Bernardo D Auria Stochastic Processes /10. Notes. Abril 13 th, 2010

Research Article Least Squares Estimators for Unit Root Processes with Locally Stationary Disturbance

Constrained Leja points and the numerical solution of the constrained energy problem

Stochastic integral. Introduction. Ito integral. References. Appendices Stochastic Calculus I. Geneviève Gauthier.

Inference for Lévy-Driven Continuous-Time ARMA Processes

A TIME SERIES PARADOX: UNIT ROOT TESTS PERFORM POORLY WHEN DATA ARE COINTEGRATED

Lecture 4: Introduction to stochastic processes and stochastic calculus

Tools of stochastic calculus

On It^o's formula for multidimensional Brownian motion

LECTURE 15 + C+F. = A 11 x 1x1 +2A 12 x 1x2 + A 22 x 2x2 + B 1 x 1 + B 2 x 2. xi y 2 = ~y 2 (x 1 ;x 2 ) x 2 = ~x 2 (y 1 ;y 2 1

Time Series: Theory and Methods

On the relation between the Smith-Wilson method and integrated Ornstein-Uhlenbeck processes

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

Linear Regression and Its Applications

ARMA Models: I VIII 1

A MODEL FOR THE LONG-TERM OPTIMAL CAPACITY LEVEL OF AN INVESTMENT PROJECT

Semiclassical analysis and a new result for Poisson - Lévy excursion measures

Automatic Local Smoothing for Spectral Density. Abstract. This article uses local polynomial techniques to t Whittle's likelihood for spectral density

Prediction of ESTSP Competition Time Series by Unscented Kalman Filter and RTS Smoother

Numerical Integration of SDEs: A Short Tutorial

2 FRED J. HICKERNELL the sample mean of the y (i) : (2) ^ 1 N The mean square error of this estimate may be written as a sum of two parts, a bias term

Temporal aggregation of stationary and nonstationary discrete-time processes

SDE Coefficients. March 4, 2008

Spurious Chaotic Solutions of Dierential. Equations. Sigitas Keras. September Department of Applied Mathematics and Theoretical Physics

Homogeneous Stochastic Differential Equations

LAN property for sde s with additive fractional noise and continuous time observation

A numerical method for solving uncertain differential equations

1. Stochastic Processes and filtrations

An Alternative to CARMA Models via Iterations of Ornstein Uhlenbeck Processes

Stochastic Differential Equations.

Man Kyu Im*, Un Cig Ji **, and Jae Hee Kim ***

COMPUTER ALGEBRA DERIVATION OF THE BIAS OF LINEAR ESTIMATORS OF AUTOREGRESSIVE MODELS

GENERALIZED ANNUITIES AND ASSURANCES, AND INTER-RELATIONSHIPS. BY LEIGH ROBERTS, M.Sc., ABSTRACT

An application of hidden Markov models to asset allocation problems

Lecture 2: Review of Prerequisites. Table of contents

OPTIMAL SOLUTIONS TO STOCHASTIC DIFFERENTIAL INCLUSIONS

On the quantiles of the Brownian motion and their hitting times.

Sequential maximum likelihood estimation for reflected Ornstein-Uhlenbeck processes

Applications of Ito s Formula

ON THE APPROXIMATION OF CONTINUOUS TIME THRESHOLD ARMA PROCESSES*

Pure Random process Pure Random Process or White Noise Process: is a random process {X t, t 0} which has: { σ 2 if k = 0 0 if k 0

K. Ito's stochastic calculus isacollection of tools which permitusto perform operations

Research Article A Necessary Characteristic Equation of Diffusion Processes Having Gaussian Marginals

Generalized Gaussian Bridges of Prediction-Invertible Processes

Asymptotical distribution free test for parameter change in a diffusion model (joint work with Y. Nishiyama) Ilia Negri

Simulating Properties of the Likelihood Ratio Test for a Unit Root in an Explosive Second Order Autoregression

Representing Gaussian Processes with Martingales

Stationary Gaussian Markov Processes As Limits of Stationary Autoregressive Time Series

STOCHASTIC DIFFERENTIAL EQUATIONS WITH EXTRA PROPERTIES H. JEROME KEISLER. Department of Mathematics. University of Wisconsin.

Some Aspects of Universal Portfolio

Maximum Likelihood Estimation

On the Goodness-of-Fit Tests for Some Continuous Time Processes

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

Nonparametric Drift Estimation for Stochastic Differential Equations

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

Some Terminology and Concepts that We will Use, But Not Emphasize (Section 6.2)

Time Series I Time Domain Methods

Time Series Analysis. Asymptotic Results for Spatial ARMA Models

On prediction and density estimation Peter McCullagh University of Chicago December 2004

Continuous Time Finance

Goodness-of-Fit Tests for Time Series Models: A Score-Marked Empirical Process Approach

Introduction to numerical simulations for Stochastic ODEs

460 HOLGER DETTE AND WILLIAM J STUDDEN order to examine how a given design behaves in the model g` with respect to the D-optimality criterion one uses

output dimension input dimension Gaussian evidence Gaussian Gaussian evidence evidence from t +1 inputs and outputs at time t x t+2 x t-1 x t+1

Regularity of the density for the stochastic heat equation

A Stable Finite Dierence Ansatz for Higher Order Dierentiation of Non-Exact. Data. Bob Anderssen and Frank de Hoog,

The Restricted Likelihood Ratio Test at the Boundary in Autoregressive Series

On the Central Limit Theorem for an ergodic Markov chain

f(z)dz = 0. P dx + Qdy = D u dx v dy + i u dy + v dx. dxdy + i x = v

Pruscha: Semiparametric Estimation in Regression Models for Point Processes based on One Realization

The Uniform Integrability of Martingales. On a Question by Alexander Cherny

On averaging principle for diusion processes with null-recurrent fast component

Clases 11-12: Integración estocástica.

Nested Uncertain Differential Equations and Its Application to Multi-factor Term Structure Model

Verona Course April Lecture 1. Review of probability

Transcription:

Continuous-time Gaussian Autoregression Peter Brockwell, Richard Davis and Yu Yang, Statistics Department, Colorado State University, Fort Collins, CO 853-877 Abstract The problem of tting continuous-time autoregressions (linear and non-linear) to closely and regularly spaced data is considered. For the linear case Jones (98) and Bergstrom (985) used state-space representations to compute exact maximum likelihood estimators and Phillips (959) did so by tting an appropriate discretetime ARMA process to the data. In this paper we use exact conditional maximum likelihood estimators for the continuously-observed process to derive approximate maximum likelihood estimators based on the closely-spaced discrete observations. We do this for both linear and non-linear autoregressions and indicate how the method can be modied also to deal with non-uniformly but closely-spaced data. Examples are given to indicate the accuracy of the procedure. AMS Mathematics Subject Classication: primary: 6M, 6H secondary: 6M9 Keywords and phrases: Continuous-time autoregression, threshold autoregression, maximum likelihood, sampled process, Wiener measure, Radon-Nikodym derivative, Cameron- Martin-Girsanov formula.

. Introduction. This paper is concerned with estimation for continuous-time Gaussian autoregressions, both linear and non-linear, based on observations made at closely-spaced times. The idea is to use the exact conditional probability density of the (p ) st derivative of an autoregression of order p with respect to Wiener measure in order to nd exact conditional maximum likelihood estimators of the parameters under the assumption that the process is observed continuously. The resulting estimates are expressed in terms of stochastic integrals which are then approximated using the available discrete-time observations. In Section we dene the continuous-time AR(p) (abbreviated to CAR(p)) process driven by Gaussian white noise and briey indicate the relation between the CAR(p) process fy (t); t g and the sampled process fy n (h) := Y (nh); n = ; ; ; : : : g. The process fy n (h) g is a discrete-time ARMA process, a result employed by Phillips (959) to obtain maximum likelihood estimates of the parameters of the continuous-time process based on observations of fy n (h) ; nh T g. From the state-space representation of the CAR(p) process it is also possible to express the likelihood of observations of fy n (h) g directly in terms of the parameters of the CAR(p) process and thereby to compute maximum likelihood estimates of the parameters as in Jones (98) and Bergstrom (985). For a CAR() process we use the asymptotic distribution of the maximum likelihood estimators of the coecients of the ARMA process fy n (h) g to derive the asymptotic distribution, as rst T! and then h!, of the estimators of the coecients of the underlying CAR process. In Section 3 we derive the probability density with respect to Wiener measure of the (p ) st derivative of the (not-necessarily linear) autoregression of order p. This forms the basis for the inference illustrated in Sections 4, 5 and 6. In the non-linear examples considered we restrict attention to continuous-time threshold autoregressive (CTAR) processes, which are continuous-time analogues of the discrete-time threshold models of Tong (983). In Section 4 we apply the results to (linear) CAR(p) processes, deriving explicit expressions for the maximum likelihood estimators of the coecients and illustrating the performance of the approximations when the results are applied to a discretely observed CAR() process. In Section 5 we consider applications to CTAR() and CTAR() processes with known threshold and in Section 6 we show how the technique can be adapted to include estimation of the threshold itself. The technique is also applied to the analysis of the Canadian lynx trappings, 8-934,. The Gaussian CAR(p) and corresponding sampled processes.

A continuous-time Gaussian autoregressive process of order p > is dened symbolically to be a stationary solution of the stochastic dierential equation, (:) a(d)y (t) = bdw (t); where a(d) = D p + a D p + + a p, the operator D denotes dierentiation with respect to t and fw (t); t g is standard Brownian motion. Since DW (t) does not exist as a random function, we give meaning to equation (.) by rewriting it in state-space form, (:) Y (t) = (b; ; : : : ; )X(t); where the state vector X(t) = (X (t); : : : ; X p (t)) T satises the It^o equation, (:3) dx(t) = AX(t)dt + edw (t); with A = 6 4....... a p a p a p a From equation (.3) we see that X j (t) is the j th 3 7 5 and e = 6 4. 3 7 5 mean-square and pathwise derivative D j X (t); j = ; : : : ; p. We are concerned in this paper with inference for the autoregressive coecients, a ; : : : ; a p, based on observations of the process Y at times ; h; h; : : : ; h[t=h], where h is small and [x] denotes the integer part of x. One approach to this problem, due to Phillips (959), is to estimate the coecients of the discrete-time ARMA process fy n (h) := Y (nh); n = ; ; ; : : : g and from these estimates to obtain estimates of the coecients a ; : : : ; a p in equation (.). The sampled process fy n (h) g is a stationary solution of the Gaussian ARMA(p ; q ) equations, (:4) (B)Y (h) n = (B)Z n ; fz n g WN(; ); where (B) and (B) are polynomials in the backward shift operator B of orders p and q respectively, where p p and q < p. (For more details see, e.g., Brockwell (995).) An alternative approach is to use equations (.) and (.3) to express the likelihood of observations of fy n (h) g directly in terms of the parameters of the CAR(p) process and then to compute numerically the maximum likelihood estimates of the parameters as in Jones (98) and Bergstrom (985). In this paper we take a dierent point of view by assuming initially that the process Y is observed continuously on the interval [; T ]. Under this assumption it is possible to 3 :

calculate exact (conditional on X()) maximum likelihood estimators of a ; : : : ; a p. To deal with the fact that observations are made only at times ; h; h; : : :, we approximate the exact solution based on continuous observations using the available discrete-time observations. This approach has the advantage that for very closely spaced observations it performs well and is extremely simple to implement. This idea can be extended to non-linear (in particular threshold) continuous-time autoregressions. We illustrate this in Sections 4, 5 and 6. The assumption of uniform spacing, which we make in all our examples, can also be relaxed providing the maximum spacing between observations is small. Before considering this alternative approach, we rst examine the method of Phillips as applied to CAR() processes. This method has the advantage of requiring only the tting of a discrete-time ARMA process to the discretely observed data and the subsequent transformation of the estimated coecients to continuous-time equivalents. We derive the asymptotic distribution of these estimators as rst T! and then h!. Example. For the CAR() process dened by the sampled process fy (h) n Y (h) n (D + a D + a )Y (t) = bdw (t); = Y (nh); n = ; ; : : : g satises (h) Y (h) n (h) Y (h) n = Z n + (h) Z n ; fz t g WN(; (h)): For xed h, as T!, the maximum likelihood estimator of = ( (h) ; (h) ; (h) ) T based on observations Y (h) ; : : : ; Y (h) satises (see Brockwell and Davis (99), p.58) [T =h] p (:5) T=h( ^ ) ) N(; M()): where (:6) M() = " EU t U T t EU t V T t EV t U T t EV t V T t ; and the random vectors U t and V t are dened as U t = (U t ; : : : ; U t+ p ) T and V t = (V t ; : : : ; V t+ q ) T, where fu t g and fv t g are stationary solutions of the autoregressive equations, (:7) (B)U t = Z t and (B)V t = Z t : In order to determine the asymptotic behaviour as T! of the maximum likelihood estimators ( ^ (h); ^ (h)), we consider the top left submatrix M of the matrix M. For small h we nd that M has the representation, (:8) M = " (a h + p 3 ( p 3)a h + 4 3 ( p 3)a 3 h 3 ) 4

+ " a a h 3 + O(h 4 ) as h! : The mapping from ( ; ) to (a ; a ) is as follows: The matrix a = log( )=h; r a = h log + 4 + C = therefore has the asymptotic expansion (:9) C = 4! log " @a @a @ @ @a @a @ @ h h + a h + a +a h + r 4 +! + a h + a h + h + a h + a 4a h + : 3 5 : From (.8) and (.9) we nd that (:) CM C T = h " a ( + o()) as h! : a a and hence, from (.5) that the maximum likelihood estimator ^a of a = (a ; a ) T based on observations of Y at times ; h; h; : : : ; h[t=h], satises p T (^a a) ) N(; V ); as T! ; where (:) V = " a ( + o()) as h! : a a Remark. Since the moving average coecient (h) of the sampled process is also a function of the parameters a and a, and hence of (h) and (h), the question arises as to whether the discrete-time likelihood maximization should be carried out subject to the constraint imposed by the functional relationship between (h) ; (h) and (h). However, as we shall see, the unconstrained estimation which we have considered in the preceding example leads to an asymptotic distribution of the estimators which, as h!, converges to that of the maximum likelihood estimators based on the process observed continuously on the interval [; T ]. This indicates, at least asymptotically, that there is no gain in using the more complicated constrained maximization of the likelihood, so that widely available standard ARMA tting techniques can be used. 5

Remark. As the spacing h converges to zero, the autoregressive roots exp( j h) converge to, leading to numerical diculties in carrying out the discrete-time maximization. For this reason we consider next an approach which uses exact results for the continuously observed process to develop approximate maximum likelihood estimators for closely-spaced discrete-time observations. The same approach can be used not only for linear continuous-time autoregressions, but also for non-linear autoregressions such as continuous-time analogues of the threshold models of Tong (983). 3. Inference for Continuously Observed Autoregressions We now consider a more general form of (.), i.e. (3:) (D p + a D p + + a p )Y (t) = b(dw (t) + c); in which we allow the parameters a ; : : : ; a p and c to be bounded measurable functions of Y (t) and assume that b is constant. In particular if we partition the real line into subintervals, ( ; y ], (y ; y ]; : : :, (y m ; ), on each of which the parameter values are constant, then we obtain a continuous-time analogue of the threshold models of Tong (983) which we shall refer to as a CTAR(p) process. Continuous-time threshold models have been used by a number of authors (e.g. Tong and Yeung (99), Brockwell and Williams (997)) for the modelling of nancial and other time series). The equation (3.) has a state space representation analogous to (.) and (.3), namely (3:) Y (t) = bx (t); where (3:3) dx = X (t)dt; dx = X (t)dt;. dx p = X p (t)dt; dx p = [ a p X (t) a X p (t) + c]dt + dw (t); and we have abbreviated a i (Y (t)) and c(y (t)) to a i and c respectively. We show next that (3.3) with initial condition X() = x = (x ; x ; ; x p ) T has a unique (in law) weak solution X = (X(t); t T ) and determine the probability density of the random function X p = (X p (t); t T ) with respect to Wiener measure. For parameterized functions a i and c, this allows the possibility of maximization of the likelihood, conditional on X() = x, of fx p (t); t T g. Of course a complete set of observations 6

of fx p (t); t T g is not generally available unless X is observed continuously. Nevertheless the parameter values which maximize the likelihood of fx p (t); t T g can be expressed in terms of observations of fy (t); t T g as described in subsequent sections. If Y is observed at discrete times, the stochastic integrals appearing in the solution for continuously observed autoregressions will be approximated by corresponding approximating sums. Other methods for dealing with the problem of estimation for continuous-time autoregressions based on discrete-time observations are considered by Stramer and Roberts (4) and by Tsai and Chan (999, ). Assuming that X() = x, we R can write X(t) in terms of fx R p (s); s tg using t the relations, X p (t) = x p + X t p (s)ds, : : :, X (t) = x + X (s)ds. The resulting functional relationship will be denoted by (3:4) X(t) = F(X p ; t): Substituting from (3.4) into the last equation in (3.3), we see that it can be written in the form, (3:5) dx p = G(X p ; t)dt + dw (t); where G(X p ; t), like F(X p ; t), depends on fx p (s); s tg. Now let B be standard Brownian motion (with B() = x p ) dened on the probability space (C[; T ]; B[; T ]; P xp ) and, for t T, let F t = fb(s); s tg _ N, where N is the sigma-algebra of P xp -null sets of B[; T ]. The equations (3:6) dz = Z dt; dz = Z dt;. dz p = Z p dt; dz p = db(t); with Z() = x = (x ; x ; ; x p ) T, clearly have the unique strong solution, Z(t) = F(B; t); where F is dened as in (3.4). Let G be the functional appearing in (3.5) and suppose that ^W is the Ito integral dened by ^W () = x p and (3:7) d ^W (t) = G(B; t)dt + db(t) = G(Z p ; t)dt + dz p (t): For each T, we now dene a new measure ^P x on F T by (3:8) d ^P x = M(B; T )dp xp ; where (3:9) M(B; T ) = exp Z T G (B; s)ds + G(B; s)dw (s) : 7

Then by the Cameron-Martin-Girsanov formula (see e.g. O= ksendal (998), p.5), f ^W (t); t T g is standard Brownian motion under ^P x. Hence we see from (3.7) that the equations (3.5) and (.3) with initial condition X() = x have, for t [; T ], the weak solutions (Z p (t); ^W (t)) and (Z(t); ^W (t)) respectively. Moreover, by Proposition 5.3. of Karatzas and Shreve (99), the weak solution is unique in law, and by Theorem.. of Stroock and Varadhan (979) it is non-explosive. If f is a bounded measurable functional on C[; T ], ^E x f(z p ) = E xp (M(B; T )f(b)) = Z f()m(; T )dp xp (): In other words, M(; T ) is the density at C[; T ], conditional on X() = x, of the distribution of X p with respect to the Wiener measure P xp and, if we observed X p =, we could compute conditional maximum likelihood estimators of the unknown parameters by maximizing M(; T ). 4. Estimation for CAR(p) Processes For the CAR(p) process dened by (.), if fx(s) = (x (s); x (s); : : : ; x p (s)) T ; s T g denotes the realized state process on the interval [; T ], we have, in the notation of Section 3, (4:) log M(x p ; s) = where G ds Gdx p (s); (4:) G = a x p (s) a x p (s) a p x (s): Dierentiating log M partially with respect to a ; : : : ; a p and setting the derivatives equal to zero gives the maximum likelihood estimators, conditional on X() = x(), (4:3) 6 4 ^a. ^a p 3 7 5 = 6 4 x p ds.... x p x ds x p x ds. x ds 3 7 5 6 4 x p dx p. x dx p Note that this expression for the maximum likelihood estimators is unchanged if x is replaced throughout by y, where y denotes the observed CAR(p) process and y j denotes its j th derivative. 3 7 5 : 8

Dierentiating log M twice with respect to the parameters a ; : : : ; a p, taking expected values and assuming that the zeroes of the autoregressive polynomial a all have negative real parts, we nd that (4:4) E @ log M @a T as T! ; where is the covariance matrix of the limit distribution as T! of the random vector (X p (t); X p (t); : : : ; X (t)) T. It is known (see Arato (98)) that (4:5) = [m ij ] p i;j= ; where m ij = m ji and for j i, m ij = 8 < : if j i is odd; P k= ( )k a i k a j+k otherwise; where a := and a j := if j > p or j <, and that the estimators given by (4.3) satisfy (4:6) p T (^a a) ) N(; ); where is given by (4.5). The asymptotic result (4.6) also holds for the Yule-Walker estimates of a as found by Hyndman (993). In the case p =, = a and when p =, is the same as the leading term in the expansion of the covariance matrix V in (.). In order to derive approximate maximum likelihood estimators for closely-spaced observations of the CAR(p) process dened by (.) we shall use the result (4.3) with the stochastic integrals replaced by approximating sums. Thus if observations are made at times ; h; h; : : :, we replace, for example, x (s) ds by h x (s)dx (s) by h X [T =h] i= (x((i + )h) x(ih)) ; X [T =h] 3 i= (x((i + )h) x(ih)) (x((i + 3)h) x((i + )h) + x((i + )h)); taking care, as in the latter example, to preserve the non-anticipating property of the integrand in the corresponding approximating sum. Example. For the CAR() process dened by (D + a D + a )Y (t) = bdw (t); 9

Table. Estimated coecients based on replicates on [; T ] of the linear CAR() process with a = :8 and a = :5 T = T =5 h Sample mean Estimated variance Sample mean Estimated variance of estimators of estimators of estimators of estimators. a.8.3585.7979.673 a.545.38.548.386. a.7864.344.777.6484 a.536.8.57.3799. a.5567.447.5465.478 a.495.9.4588.37 Table shows the result of using approximating sums for the estimators dened by (4.3) in order to estimate the coecients a and a. As expected, the variances of the estimators are reduced by a factor of approximately 5 as T increases from to 5 with h xed. As h increases with T xed, the variances actually decrease while the bias has a tendency to increase. This leads to mean squared errors which are quite close for h = : and h = :. The asymptotic covariance matrix in (4.6), based on continuously observed data, is diagonal with entries 3:6 and :8. For h = : and h = :, the variances 3:6=T and :8=T agree well with the corresponding entries in the table. 5. Estimation for CTAR(p) Processes The density derived in Section 3 is not restricted to linear continuous-time autoregressions as considered in the previous section. It applies also to non-linear autoregressions and in particular to CTAR models as dened by (3.) and (3.3). In this section we illustrate the application of the continuous-time maximum likelihood estimators and corresponding approximating sums to the estimation of coecients in CTAR() and CTAR() models. Example 3. Consider the CTAR() process dened by DY (t) + a () Y (t) = bdw (t); if Y (t) < ; with b > and a () 6= a (). We can write DY (t) + a () Y (t) = bdw (t); if Y (t) ; Y (t) = b X(t);

where dx(t) + a(x(t))x(t)dt = dw (t); and a(x) = a () if x < and a(x) = a () if x. Proceeding as in Section 4, log M is as in (4.) with (5:) G = a () x(s)i x(s)< a () x(s)i x(s) : Maximizing log M as in Section 4, we nd that and ^a () = ^a () = I x(s)<x(s)dx(s) I x(s)<x(s) ds I x(s)x(s)dx(s) I x(s)x(s) ds ; where, as in Section 4, x can be replaced by y in these expressions. For observations at times ; h; h; : : : ; with h small the integrals in these expressions were replaced by corresponding approximating sums and the resulting estimates are shown in Table. Table. Estimated coecients based on replicates on [; T ] of the threshold AR() with threshold r =, a () = 6, a () = :5 T = T =5 h Sample mean Estimated variance Sample mean Estimated variance of estimators of estimators of estimators of estimators. a () 6.45.47 5.9965.785 a ().54.484.4986.89. a () 5.8978.3947 5.847.6785 a ().535.477.4875.883. a () 4.7556.7969 4.785.456 a ().389.384.368.7 Again we see that as T increases from to 5, the variances of the estimators are reduced by a factor of approximately 5. As h increases with T xed, the variances decrease while the bias tends to increase, the net eect being (as expected) an increase in mean squared error with increasing h. Example 4. Consider the CTAR() process dened by D Y (t) + a () DY (t) + a () Y (t) = bdw (t); if Y (t) < ; D Y (t) + a () DY (t) + a () Y (t) = bdw (t); if Y (t) ;

with a () 6= a () or a () 6= a (), and b >. We can write where Y (t) = (b; ) X(t); dx(t) = AX(t)dt + e dw (t); and A = A () if x < and A = A () if x, where A () = " a () a () ; A () = " a () a () ; e = Proceeding as in Section 4, log M is as in (4.) with (5:) G = a () x (s) a () x(s) I x(s)< + a () x (s) a () x(s) " : I x(s) : Maximizing log M as in Section 4, we nd that " ^a () ^a () = " I R x(s)<x (s)ds T I x(s)<x (s)x (s)ds I x(s)<x (s)x (s)ds I x(s)<x (s)ds " I x(s)<x (s)dx (s) I x(s)<x (s)dx (s) ; while [^a () ; ^a () ] T satises the same equation with I x(s)< replaced throughout by I x(s). As in Section 4, x can be replaced by y in these expressions. For observations at times ; h; h; : : : ; with h small, the integrals in these expressions were replaced by corresponding approximating sums and the resulting estimates are shown in Table 3. The pattern of results is more complicated in this case. As T is increased from to 5 with h xed, the sample variances all decrease, but in a less regular fashion than in Tables and. As h increases with T xed, the variances also decrease. The mean squared errors for h = : and h = : are again quite close.

Table 3. Estimated coecients based on replicates on [; T ] of the threshold AR() with threshold r =, a () = :5, a () = :4, a () = 4:6, a () = T = T =5 h Sample mean Estimated variance Sample mean Estimated variance of estimators of estimators of estimators of estimators. a ().587.544.57.8 a ().4763.49.463.48 a () 4.684.4 4.5755.3995 a ().386.769.456.888. a ().56.534.563.95 a ().479.456.435.473 a () 4.389.983 4.348.3746 a ().697.6895.5.859. a ().59.477.498.85 a ().44.3489.385.4 a ().753.3.74.874 a ().7654.438.5599.5 6. Estimation when the threshold is unknown In the previous section we considered the estimation of the autoregressive coecients only, under the assumption that the threshold r is known. In this section we consider the corresponding problem when the threshold also is to be estimated. The idea is the same, that is to maximize the (conditional) likelihood of the continuously-observed process, using the closely spaced discrete observations to approximate what would be the exact maximum likelihood estimators if the continuously-observed data were available. We illustrate rst with a CTAR() process. The goal is to use observations fy(kh); k = ; ; : : : ; < kh T g, with h small, to estimate the parameters a () ; a () ; c () ; c () ; b and r in the following model. (6:) D (Y (t) r) + a () (Y (t) r) + c () = bdw t ; Y (t) < r; D (Y (t) r) + a () (Y (t) r) + c () = bdw t ; Y (t) r: The process Y = Y r, satises the threshold autoregressive equations, DY (t) + a () Y (t) + c () = bdw t ; Y (t) < ; DY (t) + a () Y (t) + c () = bdw t ; Y (t) ; 3

with state-space representation, Y (t) = bx(t); where as in equation (3.5), and G(x; s) = a () x(s) + c() b dx(t) = G(X; t)dt + dw (t);! I x(s)< a () x(s) + c() b Substituting for G in the expression (4.), we obtain log M(x(s); s) = = + a () = b + G ds! x(s) + c() I x(s)<ds + b a () x(s) + c() b a () y + c () a () y + c ()! Gdx(s) a () I x(s)< dx(s) + Iy <ds + I y <dy +! I x(s) :! x(s) + c() I x(s)ds b a () a () y + c () a () y + c () x(s) + c() b Iy ds! I y dy : I x(s) dx(s) Minimizing log M(x(s); s) with respect to a (), a (), c (), and c () with b xed gives, (6:) ^a () (r) ^c () (r) " y I y <ds I y <ds = y I y <dy I y <ds " = y I y <ds I y <ds I y <dy y I y <ds y I y <ds I y <dy y I y <ds y I y <dy y I y <ds y I y <ds ; with analogous expressions for ^a () and ^c (). An important feature of these equations is that they involve only the values of y = y r and not b. 4

For any xed value of r and observations y, we can therefore compute the maximum likelihood estimators ^a () (r); ^a () (r); ^c () (r) and ^c () (r) and the corresponding minimum value, m(r), of b log M. The maximum likelihood estimator ^r of r is the value which minimizes m(r) (this minimizing value also being independent of b). The maximum likelihood estimators of a () ; a () ; c () and c () are the values obtained from (6.) with r = ^r. Since the observed data are the discrete observations fy(h); y(h); y(3h); : : :g, the calculations just described are all carried out with the integrals in (6.) replaced by approximating sums as described in Section 4. If the data y are observed continuously, the quadratic variation of y on the interval [; T ] is exactly equal to b T. The discrete approximation to b based on fy(h); y(h); : : : g is v uut [T =h] X (6:3) ^b = k= (y((k + )h) y(kh)) =T : Example 5. Table 4 shows the results obtained when the foregoing estimation procedure is applied to a CTAR() process dened by (6.) with a () = 6, c () = :5, a () = :5, c () = :4, b = and r =. The pattern of results is again rather complicated. As expected however there is a clear reduction in sample variance of the estimators as T is increased with h xed. For T = the mean squared errors of the estimators all increase as h increases, with the mean squared errors when h = : and h = : being rather close and substantially better than those when h = :. Example 6. Although the procedure described above is primarily intended for use in the modelling of very closely spaced data, in this example we illustrate its performance when applied to the natural logarithms of the annual Canadian lynx trappings, 8-934 (see e.g. Brockwell and Davis (99), p.559). Linear and threshold autoregressions of order two were tted to this series by Tong and Yeung (99) and a linear CAR() model using a continuous-time version of the Yule-Walker equations by Hyndman (993). The threshold AR() model tted by Tong and Yeung (99) to this series was (6:4) with (6.5) D Y (t) + a () DY (t) + a () Y (t) = b DW (t); if Y (t) < r; D Y (t) + a () DY (t) + a () Y (t) = b DW (t); if Y (t) r; a () = :354; a () = :5; b = :77; 5

Table 4. The sample mean and sample variance of the estimators of the parameters of the model (6.) based on replicates of the process on [; T ]. The parameters of the simulated process are a () = 6, c () = :5, a () = :5, c () = :4, b = and r =. T = T =5 T = h Sample Sample Sample Sample Sample Sample mean variance mean variance mean variance. a () 5.979.577 5.9758.95 5.9835.94 c ().3787.778.3448.56.383.753 a ().749.45.537.5.578.4 c ().89.476.345.73.36.33 b.9996 5. 6.999 4.78 7.999 4.84 7 r 9.9963.44 9.9769.4 9.988.. a () 5.7.375 5.757.834 5.764.9 c ().47.754.335.699.3535.75 a ().7373.448.5567.538.536.39 c ().877.598.37.357.347.6 b.994 4.8 5.993 4.55 6.997 4.9 6 r..78 9.987.58 9.984.4. a () 4.66.786 4.87.587 4.5.7 c ().3953.5638.834.87.78.944 a ().738.59.594.85.585.34 c ().636.39.658.3.666.47 b.99 5. 4.98 4.6 5.96 4.79 5 r.74.45.38.9.3.86 a () = :877; a () = :47; b = :87; and threshold r = :857. An argument exactly parallel to that for the CTAR() process at the beginning of this section permits the estimation of the coecients and threshold of a CTAR() model of this form with b = b = b, h = and with time measured in years. It leads to the coecient estimates, (6.6) a () = :363; a () = :93; b = :5; a () = :5; a () = :947; b = :5; 6

with estimated threshold r = :478. (Because of the large spacing of the observations in this case it is dicult to obtain a good approximation to the quadratic variation of the derivative of the process. The coecient b was therefore estimated by a simple one-dimensional maximization of the Gaussian likelihood (GL) of the original discrete observations (computed as described by Brockwell()), with the estimated coecients xed at the values specied above.) In terms of the Gaussian likelihood of the original data, the latter model (with log(gl) = :5) is considerably better than the Tong and Yeung model (for which log(gl) = 44:4). Using our model as an initial approximation for maximizing the Gaussian likelihood of the original data, we obtain the following more general model, which has higher Gaussian likelihood than both of the preceding models ( log(gl) = 6:6). (6:7) D Y (t) + :8DY (t) + :38Y (t) :345 = :5DW (t); D Y (t) + :75DY (t) + :45Y (t) + :5 = :645DW (t); if Y (t) < :5; if Y (t) :5; Simulations of the model (6.4) with parameters as in (6.5) and (6.6) and of the model (6.7) are shown together with the logged and mean-corrected lynx data in Figure. As expected, the resemblance between the sample paths and the data appears to improve with increasing Gaussian likelihood. (a) Model (6.5) (b) Model (6.6) -6-4 - 4-6 -4-4 4 6 8 time (c) Model (6.7) 4 6 8 time (d) Lynx data -6-4 - 4-6 -4-4 4 6 8 time 4 6 8 time Figure : Figures (a) and (b) show simulations of the CTAR model (6.4) for the logged and mean-corrected lynx data when the parameters are given by (6.5) and (6.6) respectively. Figure (c) shows a simulation (with the same driving noise as in Figures (a) and (b)) of the model (6.7). Figure (d) show the logged and mean-corrected lynx series itself. 7

7. Conclusions From the Radon-Nikodym derivative with respect to Wiener measure of the distribution of the (p ) th derivative of a continuous-time linear or non-linear autoregression, observed on the interval [; T ], we have shown how to compute maximum likelihood parameter estimators, conditional on the initial state vector. For closely-spaced discrete observations, the integrals appearing in the estimators are replaced by approximating sums. The examples illustrate the accuracy of the approximations in special cases. If the observations are not uniformly spaced but the maximum spacing is small, appropriately modied approximating sums can be used in order to approximate the exact solution for the continuously observed process. Acknowledgments The authors are indebted to the National Science Foundation for support of this work under Grant DMS-389 and PB for the additional support of Deutsche Forschungsgemeinschaft, SFB 386, at Technische Universitat Munchen. We are also indebted to two referees for valuable comments and suggestions. References Arato, M. (98). Linear Stochastic Systems with Constant Coecients, Springer Lecture Notes in Control and Information Systems, 45, Springer-Verlag, Berlin. Bergstrom, A.R. (985). The estimation of parameters in non-stationary higher-order continuous-time dynamic models. Econometric Theory,, 369{385. Brockwell, P.J. and R.A. Davis (99). Time Series: Theory and Methods, nd edition. Springer-Verlag, New York. Brockwell, P.J. (995). A note on the embedding of discrete-time ARMA processes. J. Time Series Analysis 6, 45{46. Brockwell, P.J. and R.J. Williams (997). On the existence and application of continuoustime threshold autoregressions of order two. Adv. Appl. Prob. 9, 5{7. Brockwell, P.J. (). Continuous-time ARMA Processes, Handbook of Statistics 9; Stochastic Processes: Theory and Methods, D.N. Shanbhag and C.R. Rao, (eds), 49{76, Elsevier, Amsterdam. Hyndman, R.J. (993). Yule-Walker estimates for continuous-time autoregressive models. J. Time Series Analysis 4, 8{96. Jones, R.H. (98). Fitting a continuous time autoregression to discrete data. Applied Time Series Analysis II ed. D.F. Findley. Academic Press, New York, 65{68. 8

Karatzas, Ioannis and Steven E. Shreve (99). Brownian Motion and Stochastic Calculus Springer-Verlag, New York. O= ksendal, Bernt (998). Stochastic Dierential Equations, 5 th ed. Springer-Verlag, Berlin. Phillips, A.W. (959). The estimation of parameters in systems of stochastic dierential equations. Biometrika 46, 67{76. Stramer, O. and G.O. Roberts (4). On Bayesian analysis of non-linear continuoustime autoregression models. Technical Report 37, Department of Statistics and Actuarial Science, The University of Iowa, http://www.stat.uiowa.edu/techrep/. Stroock, D.W. and S.R.S. Varadhan (979). Multidimensional Diusion Processes. Springer- Verlag, Berlin. Tsai, H. and K.S. Chan (999). A new EM method for estimating continuous-time autoregressive models. Technical Report 85, Department of Statistics and Actuarial Science, The University of Iowa. Tsai, H. and K.S. Chan (). Comparison of two discretization methods for estimating continuous-time autoregressive models, in Statistics and Finance: An Interface, Wai-Sum Chan, Wai Keung Li and Howell Tong (eds.). Imperial College Press, London, 68{85. Tong, H. (983). Threshold Models in Non-linear Time Series Analysis, Springer Lecture Notes in Statistics. Springer-Verlag, New York. Tong, H. (99). Non-linear Time Series : A Dynamical System Approach. Clarendon Press, Oxford. Tong, H. and I. Yeung (99). Threshold autoregressive modelling in continuous time. Statistica Sinica, 4{43. Wilson, G.T. (969). Factorization of the generating function of a pure moving average process. SIAM J. Num. Analysis 6, {7. 9