On the Central Limit Theorem for an ergodic Markov chain

Similar documents
Stability of Cyclic Threshold and Threshold-like Autoregressive Time Series Models

THE GEOMETRICAL ERGODICITY OF NONLINEAR AUTOREGRESSIVE MODELS

Practical conditions on Markov chains for weak convergence of tail empirical processes

On moving-average models with feedback

Lecture 10. Theorem 1.1 [Ergodicity and extremality] A probability measure µ on (Ω, F) is ergodic for T if and only if it is an extremal point in M.

A regeneration proof of the central limit theorem for uniformly ergodic Markov chains

University of Toronto Department of Statistics

GEOMETRIC ERGODICITY OF NONLINEAR TIME SERIES

STA205 Probability: Week 8 R. Wolpert

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

Lectures on Stochastic Stability. Sergey FOSS. Heriot-Watt University. Lecture 4. Coupling and Harris Processes

Geometric ρ-mixing property of the interarrival times of a stationary Markovian Arrival Process

Establishing Stationarity of Time Series Models via Drift Criteria

Pointwise convergence rates and central limit theorems for kernel density estimators in linear processes

STOCHASTIC PROCESSES Basic notions

arxiv: v2 [econ.em] 16 Apr 2019

A Functional Central Limit Theorem for an ARMA(p, q) Process with Markov Switching

Existence, Uniqueness and Stability of Invariant Distributions in Continuous-Time Stochastic Models

Markov Chains. Arnoldo Frigessi Bernd Heidergott November 4, 2015

Advanced Computer Networks Lecture 2. Markov Processes

A Note on the Invertibility of Nonlinear ARMA models

Chapter 7. Markov chain background. 7.1 Finite state space

On Reparametrization and the Gibbs Sampler

Operations Research Letters. Instability of FIFO in a simple queueing system with arbitrarily low loads

Non-homogeneous random walks on a semi-infinite strip

25.1 Ergodicity and Metric Transitivity

Stability of nonlinear AR(1) time series with delay

THE VARIANCE CONSTANT FOR THE ACTUAL WAITING TIME OF THE PH/PH/1 QUEUE. By Mogens Bladt National University of Mexico

Lecture 8: The Metropolis-Hastings Algorithm

A generalization of Dobrushin coefficient

ON CONVERGENCE RATES OF GIBBS SAMPLERS FOR UNIFORM DISTRIBUTIONS

Countable state discrete time Markov Chains

Stochastic Processes. Theory for Applications. Robert G. Gallager CAMBRIDGE UNIVERSITY PRESS

A quick introduction to Markov chains and Markov chain Monte Carlo (revised version)

Limit theorems for dependent regularly varying functions of Markov chains

Strong approximation for additive functionals of geometrically ergodic Markov chains

INVARIANT PROBABILITIES FOR

Markov Chains (Part 3)

When is a Markov chain regenerative?

Duality in Stability Theory: Lyapunov function and Lyapunov measure

System Identification, Lecture 4

System Identification, Lecture 4

Central limit theorems for ergodic continuous-time Markov chains with applications to single birth processes

LECTURE 12 UNIT ROOT, WEAK CONVERGENCE, FUNCTIONAL CLT

Module 9: Stationary Processes

Markov Chains. X(t) is a Markov Process if, for arbitrary times t 1 < t 2 <... < t k < t k+1. If X(t) is discrete-valued. If X(t) is continuous-valued

Convergence rates in weighted L 1 spaces of kernel density estimators for linear processes

Random Process Lecture 1. Fundamentals of Probability

Markov processes Course note 2. Martingale problems, recurrence properties of discrete time chains.

Department of Econometrics and Business Statistics

The Equivalence of Ergodicity and Weak Mixing for Infinitely Divisible Processes1

Transience: Whereas a finite closed communication class must be recurrent, an infinite closed communication class can be transient:

STA 624 Practice Exam 2 Applied Stochastic Processes Spring, 2008

PIGGYBACKING THRESHOLD PROCESSES WITH A FINITE STATE MARKOV CHAIN

Chapter 2: Markov Chains and Queues in Discrete Time

11. Further Issues in Using OLS with TS Data

Stability and Rare Events in Stochastic Models Sergey Foss Heriot-Watt University, Edinburgh and Institute of Mathematics, Novosibirsk

Geometric ergodicity of the Bayesian lasso

Characterization of Information Channels for Asymptotic Mean Stationarity and Stochastic Stability of Nonstationary/Unstable Linear Systems

Lecture 5. If we interpret the index n 0 as time, then a Markov chain simply requires that the future depends only on the present and not on the past.

Deterministic. Deterministic data are those can be described by an explicit mathematical relationship

Exact Simulation of the Stationary Distribution of M/G/c Queues

Regenerative Processes. Maria Vlasiou. June 25, 2018

SIMILAR MARKOV CHAINS

Bayesian Methods with Monte Carlo Markov Chains II

STABILIZATION OF AN OVERLOADED QUEUEING NETWORK USING MEASUREMENT-BASED ADMISSION CONTROL

Positive and null recurrent-branching Process

Stochastic process. X, a series of random variables indexed by t

Local consistency of Markov chain Monte Carlo methods

Extended Renovation Theory and Limit Theorems for Stochastic Ordered Graphs

On the Applicability of Regenerative Simulation in Markov Chain Monte Carlo

Gradient Estimation for Attractor Networks

Complete moment convergence of weighted sums for processes under asymptotically almost negatively associated assumptions

Northwestern University Department of Electrical Engineering and Computer Science

1 Kernels Definitions Operations Finite State Space Regular Conditional Probabilities... 4

An almost sure invariance principle for additive functionals of Markov chains

On Differentiability of Average Cost in Parameterized Markov Chains

CONTROL SYSTEMS, ROBOTICS AND AUTOMATION Vol. XI Stochastic Stability - H.J. Kushner

Faithful couplings of Markov chains: now equals forever

Some Results on the Ergodicity of Adaptive MCMC Algorithms

Probability on a Riemannian Manifold

An Introduction to Entropy and Subshifts of. Finite Type

GAUSSIAN PROCESSES; KOLMOGOROV-CHENTSOV THEOREM

1 Introduction In this work we pursue a new family of models for time series, substantially extending, but strongly related to and based on the classi

Markov chain Monte Carlo

Lecture 11: Introduction to Markov Chains. Copyright G. Caire (Sample Lectures) 321

PRECISE ASYMPTOTIC IN THE LAW OF THE ITERATED LOGARITHM AND COMPLETE CONVERGENCE FOR U-STATISTICS

6 Markov Chain Monte Carlo (MCMC)

Introduction Random recurrence equations have been used in numerous elds of applied probability. We refer for instance to Kesten (973), Vervaat (979)

FOURIER TRANSFORMS OF STATIONARY PROCESSES 1. k=1

Generalized Resolvents and Harris Recurrence of Markov Processes

THE NOISY VETO-VOTER MODEL: A RECURSIVE DISTRIBUTIONAL EQUATION ON [0,1] April 28, 2007

Discrete Probability Refresher

A Comparison between Linear and Nonlinear Forecasts for Nonlinear AR Models

Let (Ω, F) be a measureable space. A filtration in discrete time is a sequence of. F s F t

Asymptotics and Simulation of Heavy-Tailed Processes

ELECTRICAL NETWORK THEORY AND RECURRENCE IN DISCRETE RANDOM WALKS

arxiv: v1 [math.st] 12 Aug 2017

Stationarity of Generalized Autoregressive Moving Average Models

Stochastic process for macro

Transcription:

Stochastic Processes and their Applications 47 ( 1993) 113-117 North-Holland 113 On the Central Limit Theorem for an ergodic Markov chain K.S. Chan Department of Statistics and Actuarial Science, The University of Iowa, Iowa City, USA Received 17 February 1992 Revised 15 September 1992 A simple sufficient condition for the Central Limit Theorem for functionals of Harris ergodic Markov chains is derived. The result is illustrated with an example taken from non-linear time series analysis. central limit theorem * drift criterion * ergodic Markov chain * non-linear time series model 1. Introduction Many time series models used in statistical analysis have Markovian representations so that the observations are time-invariant functionals of some possibly unobservable Markov chains. When a time series model has a Markovian representation, its probabilistic properties can be analyzed via the underlying Markov chain. Recently, the theory of Harris ergodic chains has.arisen as a new tool to analyze the stationarity of certain time series models. See Tong (1990). Here, we focus on the problem of the CLT for these Harris ergodic models. In Theorem 7.6 of Nummelin (1984), a sufficient condition for the CLT for a Harris ergodic Markov chain is stated in terms of so-called!-regularity. However, this condition is difficult to verify in practice. Our purpose is to present a criterion in terms of a drift condition (Lyapunov-Foster-Tweedie condition) for the! regularity and to state a version of the CL T using this criterion. Section 2 contains the main result. In Section 3, we illustrate the criterion via a class of non-linear time series models, the SETAR(2; 1, 1) models. 2. ~ain result We first introduce some notations and definitions. Let (Xn) be a Markov chain taking values on a measurable space (E, ~) which is assumed to be countably Correspondence to: Dr K.S. Chan, Department of Statistics and Actuarial Sciences, The University of Iowa, Iowa City, la 52242, USA. Research was partially supported by the NSF grants DMS 90-06464 and DMS 91-18626. I would like to thank Professor Howell Tong and an anonymous referee for much useful ad vices on the style of writing. 0304-4149/93/$06.00 1993-Eisevier Science Publishers B.V. All rights reserved

114 K.S. Chan/ Markov chain CLT generated. In practice, E is often taken as some Euclidean space and 'if: consists of all the Borel sets. (Xn) is (Harris) ergodic if there exists a unique invariant probability measure, say 7T, such that llr(x, )- 7TII ~ 0 as n ~ oo where Pm(x, ) is the mth step transition probability for (Xn) and 11-11 denotes the total variation norm. Henceforth, it is assumed that ( Xn) is ergodic and 7T denotes its invariant probability measure. Let 'if:+= {A E 'if:: 7r(A) > 0}. A function K: Ex 'if:~~+ is said to be a kernel if VA E 'if:, K (,A) is a measurable function, and Vx E E, K (x, )is a (non-negative) measure. The transition probability P(x, A) is a kernel. For any BE 'if:, 1 8 denotes the kernel such that I 8 (x, A)= 1 iff x E An B. For any two kernels K 1 and K 2, the product K 1K 2 is the kernel defined by Define K 1K 2(x, A)= f K 2(y, A)K 1(x, dy). where Ex( ) denotes taking expectation with the initial distribution concentrated at x and S8 ( T8 ) is the first i > 0 (i ~ 0) that X; enters B. Let f be a non-negative 7T-integrable function. A E 'if:+ is said to be /-regular if VB E 'if:+, U 8f and f are bounded over A. A non-negative measure A on E is /-regular if VB E 'if:+, Aj ~r Jf( x) A ( dx) and AU 8f are finite. A set C E 'if:+ is said to be small if 3m a positive integer, {3 a constant and v a probability measure such that, Vx E C, pm (x, ) ""'{3v( ). A set BE 'if: is said to be special if E is 1 8 -regular where 1 8 is the characteristic function of B. It is shown in Nummelin (1984) that every small set is special. Further characterizations of /-regularity are given by Nummelin (1984). We now state the main result. Theorem 1. Let (X,) be a Harris ergodic Markov chain whose invariant probability measure is denoted by 7T. Let Y, = h(x,) and h is a measurable function. It is also assumed that E ( Y,) = 0. Suppose 3 'Y > 0, A a small set, and a 7T-square-integrable and non-negative function g(x) such that g(x) and E (g(xn+l) I Xn = x) are bounded over A, g(x) ~ yih(x)l, x E A, g(x) ~ E(g(Xn+I)IXn = x) + 'Yih(x)l, X~ A. (2.1) (2.2) (2.3) Then, under any initial distribution A for X 0, the normalized partial sum I~~ 1 h (X;) I v'n is asymptotically N(O, a- 2 ) with a- 2 being finite. Ifthe initial distribution A ofx 0 is such that EA (g 2 (Xo)) < oo, then a- 2 = limn_oo EA ci;~ I h(x;) ) 2 In where EA (.) denotes taking expectation with A being the distribution of X 0 Proof. For simplicity, we write f for lhl and, with no loss of generality, we also assume that 'Y in the above conditions is equal to 1.

K.S. Chan I Markov chain CLT 115 To prove the asymptotic normality result, Theorem 7.6 in Nummelin (1984) shows that it suffices to check that the measure v( dx) = f(x) 1r( dx) is!-regular. This is done as follows. It follows from conditions (2.2) and (2.3) and the Balayage Theorem (Theorem 3.1 in Nummelin, 1984) that (2.4) Let B be any Borel set with positive 1r-measure. By formula (3.18) in Nummelin (1984), (2.5) It follows from (2.1) and (2.4) that IA\B UAf~ k 11A for some positive constant k 1 Because A is small and hence a special set, U8 1A is a bounded function. Hence, there exists a constant k 2 such that f (Uaf)(x)f(x)1T(dx)~ f g(x)j(x)1r(dx)+k2 f j(x)1r(dx). (2.6) It follows from (2.2), (2.3) and the 1r-square-integrability of g that the right-hand side of (2.6) is finite. Hence, v( dx) is!-regular. A formula for r:r 2 is given in Nummelin (1984, p. 134). We now prove the last part of the theorem on the alternative expression for r:r 2 We first assume that (X") has an atom, i.e., VxE a, P(X" EAIXn_ 1 =x) is the same and 1r(a)>O. Let T"(i) be the epoch that a is visited by (X,) for the ith time. Define l(n) to be the last epoch ~ n that X; visits a. Then, n TaAn /{n)-1 n l: h(x;) = l: h(x;) + l: ((i) + L: h(x;), (2.7) i=o i=o i=l where I" s =min(/, s) and ((i) = L:r:~~:;+ 1 h(~). It is shown in the proof of Theorem 7.6 in Nummelin (1984) that {((j), j = 1, 2, 3,... } is i.i.d. with zero mean and finite variance r:r~. Furthermore, r:r 2 of the limiting Normal distribution is equal to 1r (a) - 1 r:r~ and I ( n) In ~ 1r (a) a.s. as n ~ oo. Hence, to prove the stated form for r:r 2, it suffices to show that the variances of the first and the third terms on the right-hand side of (2.7) are o(n). Now, ( TAn )2 EA i~o h(x;) ~ EA ( Gaf?. (2.8) By (2.1)-(2.5), GJ~ a+ bg for some constants a and b. Hence, the above expectation is finite. Next, it can be shown that (2.9) n ~ L E"((U"f) 2 I(S"?;m))/n. (2.10) m=l

116 K.S. Chan I Markov chain CLT Again, by (2.1)-(2.5), 3 constants a, b such that U,J~ a+ bg. Since E,.(g 2 ) is finite, E,.(( U,.f? I(S,. ~ m)) tends to zero as m -HXJ. Hence, the right-hand side of (2.10) tends to zero as n ~ oo. In the general case, the chain can be 'split' to create an atom and then the result can be proved similarly. This completes the proof of Theorem 1. D Remarks. (l) (X,) is 1ft-irreducible if Vx and B with rp(b)>o, Prob(XnEB for some n~oix0 =x) > 0. Suppose E is some Euclidean space with C consisting of the Borel sets. It is known that if (X,) is 1ft-irreducible and Feller-continuous, i.e., V bounded and continuous function w(x), E(w(Xn+d IXn=X) is continuous in x, then any compact set with non-zero >-measure is small. See Feigin and Tweedie (1985 ). (2) Suppose (X,) is only known to be aperiodic and -irreducible and conditions (2.1)-(2.3) hold for some non-negative measurable g(x) and h(x) = 1, then it is known that (X,) is ergodic. The modified condition is known as the 'drift' criterion. If the 'drift' criterion is strengthened by replacing h(x) in (2.2) by 1 and h(x) in (2.3) by g(x), then (X,) is geometrically ergodic. See Tweedie (1983a). Under the latter condition, Tweedie (1983b) further showed that E(g(X,)) < oo. ( 3) Let Y; be the covariance of h (Xi) and h (Xi +i) under the stationary distribution. Suppose that I~o IY;I <oo. Then the formula of a 2 implies that a 2 =y 0 +2 I~i y;. (4) From the proof of the theorem, it can be seen that conditions (2.1)-(2.3) imply that A is!hi-regular. It can be similarly shown that a u-finite measure, A, is!hi-regular if, in addition to (2.1)-(2.3), J g(x)a(dx) <oo. See Chan (1991) for a detailed proof. 3. Example Consider the following simple SETAR(2; 1, 1) model: X,={X,_i+c/>10+a,, ~f X,_ 1 ~0, X,_ 1 - c/> 20+ a,, 1f X,_ 1 > 0, (3.1) where (a,) is i.i.d. N(O, 1) and a, is independent of X,_~> X,_ 2,, X 0 Here, c/> 10 and c/>20 are two positive constants. The above model is a special case of the SETAR model which is an interesting class of non-linear time series models. The SETAR model is capable of exhibiting many non-linear phenomena such as limit cycle, strange attractor,..., etc. See Tong (1990). It is shown in Chan et al. (1985) that (X,) satisfying (1.1) is ergodic and compact sets with positive Lebesgue measure are small. We now show that the CLT holds for (X,). Here, h(x) = x- JL where JL = E (X,). We try g(x) = x 2 + 1 and K = [-r, r] for some r > 0 which is to be determined later. Now, for x~o. (3.2) Together with a similar result for the case when x > 0, conditions (2.1)-(2.3) can be verified to hold for some 'Y sufficiently small and some r sufficiently large. It

K.S. Chan/ Markov chain CLT 117 remains to show that g(x) is square integrable w.r.t. the invariant distribution. This can be done by invoking point (2) of the Remarks in Section 2 and consider, e.g., x 6 + 1. We note that the Normality assumption for the noise can be substantially weakened. For example, the CLT would continue to hold if the noise has density positive everywhere and has finite sixth moment. The above example demonstrates that Theorem 1 can be effectively used to establish the validity of the CLT for functionals of an ergodic Markov chain. For further applications to non-linear time series models, see Chan (1991). References K.S. Chan, Some recurrence properties of an ergodic Markov chain with applications to CLT for functionals of non-linear time series, Tech. Rep. No. 295, Dept. of Statist., Univ. of Chicago. (Chicago, IL, 1991). K.S. Chan, J.D. Petruccelli, H. Tong and S.W. Woolford, A multiple-threshold AR(1) model, J. Appl. Probab. 22 (1985) 267-279. P.O. Feigin and R.L. Tweedie, Random coefficient autoregressive processes- a Markov chain analysis of stationarity and finiteness of moments, J. Time Ser. Anal. 6 (1985) 1-14. E. Nummelin, General Irreducible Markov Chains and Non-negative Operators (Cambridge Univ. Press, Cambridge, 1984). R.L. Tweedie, Criteria for rates of convergence of Markov chains, with applications to queueing and storage theory, in: J.F.C. Kingman and G.E.H. Reuter, eds., Probability, Statistics and Analysis (Cambridge Univ. Press, Cambridge, 1983a) pp. 260-276. R.L. Tweedie, The existence of moments for stationary Markov chains, J. Appl. Probab. 20 (1983b) 191-196. H. Tong, Non-linear Time Series: A Dynamical System Approach (Clarendon Press, Oxford, 1990).