Some Stochastic Shape Applications in. Time Series and Markov Chains. Ying Zhao. (Under the direction of Robert Lund) Abstract

Size: px
Start display at page:

Download "Some Stochastic Shape Applications in. Time Series and Markov Chains. Ying Zhao. (Under the direction of Robert Lund) Abstract"

Transcription

1 Some Stochastic Shape Applications in Time Series and Markov Chains by Ying Zhao (Under the direction of Robert Lund) Abstract This dissertation explores several shape ordering applications in statistics, primarily in time series and reversible Markov chains. The dissertation has two major goals. First, we introduce shape orderings for stationary time series and explore their convergence rate ramifications. The shapes explored include increasing likelihood ratio, decreasing hazard rate and new better than used, structures reminiscent from stochastic processes settings. Examples of ARMA(p, q) time series having these shapes are presented. The shapes are then applied to obtain explicit geometric convergence rates of several one-step-ahead forecasting quantities. The second goal of this dissertation identifies a monotonicity property in reversible Markov chains and examine consequences of this structure. In particular, we show that the return times to every state in a reversible chain have a decreasing hazard rate structure on the even time indices. Good and sometimes even optimal convergence rates of a time reversible Markov chain are deduced from this monotonicity. Index words: Autocorrelation; Convergence Rate; Decreasing Hazard Rate; Decreasing Likelihood Ratio; Increasing Likelihood Ratio; Markov Chain; Mean Squared Error; Monotonicity; New Better than Used; Partial Autocorrelation; Renewal Sequence.

2 Some Stochastic Shape Applications in Time Series and Markov Chains by Ying Zhao B.S., Tianjin University, P.R. China, 2000 M.A., York University, Canada, 2001 A Dissertation Submitted to the Graduate Faculty of The University of Georgia in Partial Fulfillment of the Requirements for the Degree Doctor of Philosophy Athens, Georgia 2005

3 c 2005 Ying Zhao All Rights Reserved

4 Some Stochastic Shape Applications in Time Series and Markov Chains by Ying Zhao Approved: Major Professor: Robert Lund Committee: Nicole Lazar William P. McCormick Jaxk Reeves John Stufken Electronic Version Approved: Maureen Grasso Dean of the Graduate School The University of Georgia August 2005

5 Acknowledgements While the work with this dissertation has been extensive and trying, it has also been exciting, instructive, and fun. I am greatly indebted to many persons; without their trust, help, support and encouragement, this dissertation would not have been possible. First of all, I would like to thank my supervisor Dr. Robert Lund for his encouraging and inspiring way that guided me to a deeper understanding of knowledge during the work. Dr. Lund has such a kind spirit as well as a wealth of knowledge and an amazing ability to effectively mentor students. Your technical and editorial advice was essential in completing this dissertation and has taught me innumerable lessons and insights on the workings of academic research in general. I am also grateful to the faculty, staff, and graduate students at the Department of Statistics, with whom I have had valuable associations. You all helped me in various ways and made my experience at University of Georgia unforgettable. Special thanks go to Dr. Nicole Lazar, Dr. William McCormick, Dr. Jaxk Reeves, Dr. John Stufken, my advisory committee members, for timely assistance and insightful feedback. I thank my family in particular my parents Kehua and Xiuling and my sister Lin for their unconditional love and endless support to pursue my interests, for always being there for me. At last, to Rui Dai thanks for supporting me with your love and understanding, all the interesting discussions concerning with the work and the happiness you brought to me in the past years of my Ph.D. study life. iv

6 Table of Contents Page Acknowledgements iv Chapter 1 Introduction References Literature Review Time Series Overview Distribution Classes of Discrete Random Variables Monotone Markov Chains Renewal Theory References Shape Orderings for Stationary Time Series Introduction Definition of Orderings Examples Convergence Rates References A Monotonicity of Reversible Markov Chains Introduction Background Results v

7 vi 4.4 Convergence Rates of Reversible Chains Examples Proofs References Conclusions and Future Work Convergence of AR(p) Coefficients How Fast Can A Time-reversible Markov Chain Converge? References

8 Chapter 1 Introduction This dissertation explores applications of shape orderings in stationary time series and reversible Markov chains. The dissertation has two major goals: (1) to introduce shape orderings for stationary time series autocovariances and explore some of their convergence rate ramifications; and (2) to identify a monotonicity property in all reversible Markov chains and examine some consequences of this structure. The first part of this research introduces stochastic shape orderings into stationary time series analyses. We suggest orderings for the autocorrelation (ACF) and partial autocorrelation (PACF) functions of stationary series and study ramifications of such structures. The orderings proposed have direct analogies in stochastic process and reliability settings and include new worse than used, new better than used, increasing hazard rate, decreasing hazard rate, increasing likelihood ratio, and decreasing likelihood ratio. Whereas these orderings are not exhaustive decreasing reversed hazard rate, new better than used in expectation, increasing hazard rate average, for example, are useful orderings that are not expounded upon (cf. Shaked and Shanthikumar 1987, 1994; Kijima 1997) the flavor of what can be achieved will become clear. The primary contribution of the first part is the introduction of stochastic orderings into the time series analyst s toolbox. We are unaware of any previous literature with this slant. Given the utility of stochastic orderings in stochastic processes and reliability, the raw idea is expected to prove fruitful. Nonetheless, we do not make 1

9 2 any attempt at completing the issue here; indeed, the applications pursued are basically limited to explicit convergence rate bounds for mean squared prediction errors and Innovations Algorithm coefficients. It is expected that such shape orderings will eventually prove useful in likelihood computations, forecasting, and quantifying closeness of weighted and least squares estimators. The second application of stochastic shape orderings in this dissertation establishes a decreasing hazard rate (DHR) structure in reversible Markov chains on countable state spaces and studies some ramifications of this property. Time reversible Markov chains arise frequently in practice (cf. Ross 1996, Stroock 2005, Chen 2005) and include many Markov chain Monte Carlo (MCMC) generated chains. The research here is motivated by the Markov chain convergence rate problem. MCMC simulation algorithms were voted as one of the top 10 algorithms of all time by the IEEE society, the official bookkeepers of algorithms. Statisticians have been clever in constructing Markov chains that converge statistically to an equilibrium distribution with preset characteristics. This allows one to study many complex systems via simulation. Unfortunately, the number of iterations n needed to run the chain for a burn-in period to reach equilibrium remains unclear. In complex simulation settings where each iteration could take days, this is not feasible even with the best available computing. Such convergence rate issues have plagued statisticians in a variety of settings. Recent progress on convergence rates for stochastically ordered Markov Chains has been made (cf. Meyn and Tweedie 1993, 1994; Lund and Tweedie 1996; Lund et al. 1996; Kijima 1997, and the references therein). Hence, one is motivated to study renewal convergence rates under more general ordering structures. Our work here shows that the return time distribution to each and every state in a countable state-space reversible Markov chain has the decreasing hazard rate (DHR) property when examined on the even time indices. This structure is first proven for finite

10 3 state Markov chains and then extended to countable state spaces via a truncation argument. The DHR property identified is then used to derive a clean and explicit convergence rate bound for reversible Markov chains. This bound is even optimal in some cases. The DHR result imparts a geometry into reversible chain analyses. For example, a DHR first return distribution implies that if a fixed state has not been visited in the last l time units, then the chances of visiting the state in the next l time units are even smaller (compared to unconditional information). Markov chain analysts have frequently used stochastic orderings to drive discourse. For instance, stochastic monotonicity was used by Lund and Tweedie (1996) in coupling arguments to extract very sharp rates of convergence for ordered chains. Lindvall (1992) and Kijima (1997) are prominent references where stochastic orderings are used to assess stability of various chains. Keilson and Kester (1978), Brown (1980), Shaked and Shanthikumar (1994), Liggett (1989), Hansen and Frenk (1991), and Berenhaut and Lund (2002) are other authors who mix renewal theory, Markov chains, and stochastic orderings. That the DHR structure is identified along the subsequence of even integers is also noteworthy. For this implies that the two-step chain might be more amenable to analysis in reversible settings than the one-step chain. For MCMC simulators, this merely involves iterating the chain twice at each step instead of once, a straightforward task. The remainder of the dissertation proceeds as follows. Chapter 2 presents a brief review of stationary time series, distribution classes of discrete random variables, stochastically monotone Markov chains, and renewal theory. Chapter 3 and 4 are selfcontained manuscripts; each chapter includes its own list of references. In Chapter 3, we clarify our definition of orderings for the autocorrelation (ACF) and partial autocorrelation (PACF) functions of stationary time series. Examples of common autoregressive moving-average (ARMA(p, q)) series which have one or more of the

11 4 introduced orderings are provided. Chapter 3 then applies the results to derive some explicit geometric decay rates for quantities encountered in forecasting of stationary series. Chapter 4 identifies a monotonicity inherent to all reversible Markov chains and examines some consequences of this structure. In particular, it shows that the return times to every state in a reversible chain have a decreasing hazard rate on the even time indices. Good, and sometimes even optimal convergence rates of a time reversible Markov chain are deduced from this monotonicity. Chapter 5 concludes the dissertation by presenting some avenues for future work. 1.1 References [1] Berenhaut, K. S., and Lund, R. B. (2002). Renewal convergence rates for DHR and NWU lifetimes, Probability in the Engineering and Informational Sciences, 16, [2] Brown, M. (1980). Bounds, inequalities, and monotonicity properties for some specialized renewal processes, The Annals of Probability, 8, [3] Chen, M. F. (2005). Eigenvalues, Inequalities and Ergodic Theory, London: Springer. [4] Hansen, B. G., and Frenk, J. B. G. (1991). Some monotonicity properties of the delayed renewal function, Journal of Applied Probability, 28, [5] Keilson, J., and Kester, A. (1978). Unimodality preservation in Markov chains, Stochastic Processes and their Applications, 7, [6] Kijima, M. (1997). Markov Processes for Stochastic Modeling, London: Chapman and Hall.

12 5 [7] Liggett, T. (1989). Total positivity and renewal theory, In: Probability, Statistics, and Mathematics, Edited by T. W. Anderson, K. B. Athreya, and D. L. Iglehart, , Boston: Academic Press. [8] Lindvall, T. (1992). Lectures on the Coupling Method. New York: Wiley. [9] Lund, R. B., Meyn, S. P., and Tweedie, R. L. (1996). Computable exponential convergence rates for stochastically ordered Markov processes, Annals of Applied Probability, 6, [10] Lund, R. B. and Tweedie, R. L. (1996). Geometric convergence rates for stochastically ordered Markov chains, Mathematics of Operations Research, 20, [11] Meyn, S. P. and Tweedie, R. L. (1993). Markov Chains and Stochastic Stability, New York: Springer. [12] Meyn, S. P. and Tweedie, R. L. (1994). Computable bounds for geometric convergence rates of Markov chains, Annals of Applied Probability, 4, [13] Ross, S. M. (1996). Stochastic Processes, second edition, New York: Wiley. [14] Shaked, M., and Shanthikumar, J. G. (1987). IFRA properties of some Markov processes with a general state space, Mathematics of Operations Research, 12, [15] Shaked, M., and Shanthikumar, J. G. (1994). Stochastic Orders and their Applications, New York: Academic Press.

13 [16] Stroock, D. (2005). An Introduction to Markov Processes, New York: Springer. 6

14 Chapter 2 Literature Review 2.1 Time Series Overview This subsection reviews some results on time series. A good general reference is Brockwell and Davis (1991). A time series is a random sequence {X t }, observed in a time-ordered fashion, with t denoting time. Here, t runs over a suitable index set t T, very often with T = {0, ±1, ±2,...}, {1, 2, 3,...}, [0, ) or (, ). In this dissertation, T will be a subset of R. Moreover, we concentrate on discrete time where the observations are taken over a discrete uniformly spaced set, as is the case when observations are reordered at fixed time intervals. Time series, unlike many other branches of statistics, relies critically on the assumption that data values represent consecutive measurements taken at equally spaced time intervals Stationary Time Series A second-order description of {X t } specifies E[X t ] and Cov(X t,x s ) for all t and s. To gain insight into the dependence between the observations in a time series, we introduce the autocovariance function, which extends the idea of a covariance matrix to infinite collections of random variables. Definition (The Autocovariance Function). If {X t,t T } is a process such that Var(X t )< for each t T, then the autocovariance function γ X (, ) of {X t } 7

15 8 is defined by γ X (r,s) = Cov(X r,x s ) = E[(X r EX r )(X s EX s )], r,s T. (2.1.1) Definition (Stationarity). The time series {X t } is said to be stationary if (i) E[X t ] 2 < for all t Z, (ii) E[X t ] is a constant for all t Z, and (iii) γ X (r,s) = γ X (r + t,s + t) for all r,s,t Z. Remark 1. The above notion is frequently referred to in the literature as weak stationarity, covariance stationarity, stationarity in the wide sense, or second-order stationarity (cf. Brockwell and Davis 1991). Remark 2. If {X t } is stationary, then γ X (r,s) = γ X (r s, 0) for all r,s Z. It is therefore convenient to redefine the autocovariance function of a stationary sequence in just one variable: γ X (h) := γ X (h, 0) = Cov(X t+h,x t ) for all t,h Z. (2.1.2) The function γ X ( ) will be referred to as the autocovariance function (ACVF) of {X t } and γ X (h) as its value at lag h. The autocorrelation function (ACF) of {X t } is defined at lag h by ρ X (h) := γ X (h)/γ X (0) = Corr(X t+h,x t ) for all t,h Z. (2.1.3) The partial autocorrelation function, like the autocorrelation function, conveys vital information regarding the dependence structure of a stationary sequence. Like the autocorrelation function, the PACF only depends on the second order properties of {X t }. The partial autocorrelation α(h) at lag h is defined as the correlation between X 1 and X h+1 adjusted for the intervening observations X 2,...,X h.

16 Definition (Partial Autocorrelation Function). The partial autocorrelation function α( ) of a stationary time series {X t } is defined at lags h 1 by 9 α(1) = Corr(X 2,X 1 ) = ρ(1), and for h 2, ( ) α(h) = Corr X h+1 P sp{1,x2,...,x h }(X h+1 ),X 1 P sp{1,x2,...,x h }(X 1 ).(2.1.4) Here, sp( ) denotes closed linear span; the projections P sp{1,x2,...,x h }(X h+1 ) and P sp{1,x2,...,x h }(X 1 ) can be calculated from the classical prediction equations: n P sp{1,z1,...,z n}(x) = α i Z i, (2.1.5) where α 0 = E[X t ], Z 0 = 1, Z 1,...,Z n are mean zero random variables and α 1,...,α n satisfy the linear system of equations i=0 n α i Cov(Z i,z j ) = Cov(X,Z j ), j = 0, 1,...,n. (2.1.6) i=0 It is perhaps less well known that the ACF and PACF essentially uniquely determine each other (Ramsey, 1974). In what follows, we let {X t } be a zero mean covariance stationary series and let γ(h) = Cov(X t+h,x t ), ρ(h) = Corr(X t+h,x t ), and α(h) = Corr(X h+1,x 1 X 2,X 3,...,X h ) denote the autocovariance (ACVF), autocorrelation (ACF), and partial autocorrelation (PACF) of {X t }, respectively, at lag h Stationary ARMA Process The family of autoregressive moving-average (ARMA) processes plays a key role in modelling time series data. The ARMA family of stationary time series is defined in terms of solutions to linear difference equations with constant coefficients. The linear structure of ARMA processes leads also to a very simple theory of linear prediction,

17 10 and ARMA models are known to be dense in stationary short memory structures (cf. Brockwell and Davis 1991). Definition (The ARMA(p, q) Processes). The sequence {X t,t = 0, ±1, ±2,...} is said to be an ARMA(p, q) process if {X t } is stationary and if for every t, X t φ 1 X t 1 φ p X t p = Z t + θ 1 Z t θ q Z t q, (2.1.7) where {Z t } WN(0,σ 2 ), where WN(0,σ 2 ) denotes a white noise sequence (uncorrelated random variables with the constant variance σ 2 ). We say that {X t } is an ARMA(p, q) process with mean µ if {X t µ} is an ARMA(p, q) process. In (2.1.7), φ 1,...,φ p are called autoregressive coefficients and θ 1,...,θ q are called movingaverage coefficients. Equation (2.1.7) can be written symbolically in the compact form φ(b)x t = θ(b)z t, t = 0, ±1, ±2,..., (2.1.8) where φ and θ are the p th and q th degree polynomials φ(z) = 1 φ 1 z φ p z p (2.1.9) and θ(z) = 1 + θ 1 z + + θ q z q, (2.1.10) and B is the backward shift operator defined by B j X t = X t j, j = 0, ±1, ±2,... (2.1.11) The polynomials φ and θ will be referred to as the autoregressive and moving-average polynomials respectively in the difference equation (2.1.8). Example (MA(q) Processes). If φ(z) 1 then

18 11 X t = θ(b)z t (2.1.12) and {X t } is said to be a moving-average sequence of order q (or MA(q)). All MAs are stationary, since they are clearly linear combinations of stationary series (in this case, white noise). Furthermore, the moving-average ACVF can be derived as γ(h) = q h σ 2 j=0 θ j θ j+ h, if h q 0, if h > q. (2.1.13) Example (AR(p) Processes). If θ(z) 1, then φ(b)x t = Z t (2.1.14) and the sequence {X t } is said to be an autoregressive sequence of order p (or AR(p)). Not every AR equation has a stationary solution; indeed, an AR(1) with φ 1 = 1 is a random walk (without an initial condition, e.g. X 0 = 0) which is not stationary. In AR(p) cases the existence and uniqueness of a stationary solution to (2.1.14) merits closer investigation. For an AR(1) model, iterating k times yields X t = Z t + φ 1 X t 1 (2.1.15) As X t = Z t + φ 1 Z t φ k 1Z t k + φ k+1 1 X t k 1. (2.1.16) φ j 1Z t j is mean-square convergent (by the Cauchy criterion) when φ 1 < 1, j=0 we conclude that the solution to (2.1.15) is X t = φ j 1Z t j. (2.1.17) j=0

19 12 and Cov(X t+h,x t ) = lim E n = σ 2 φ h 1 [ ( n j=0 j=0 φ j 1Z t+h j)( n k=0 φ 2j 1 φ k 1Z t+k ) ] = σ 2 φ h 1 /(1 φ 2 1). (2.1.18) Moreover, {X t } as defined by (2.1.17) satisfies the difference equations (2.1.16) and can be shown to be the unique (in mean square) stationary solution. Definition (Causality). An ARMA(p, q) sequence {X t } is said to be causal with respect to {Z t }, if there exists a sequence of constants {ψ j } j=0 such that ψ j < and j=0 X t = ψ j Z t j, t = 0, ±1, ±2,... (2.1.19) j=0 Definition (Invertibility). An ARMA(p, q) sequence {X t } is said to be invertible with respect to {Z t }, if there exists a sequence of constants {π j } j=0 such that π j < and j=0 Z t = π j X t j, t = 0, ±1, ±2,... (2.1.20) j=0 A fundamental ARMA result is contained in the following theorem. Theorem Suppose that {X t } satisfies the ARMA(p,q) difference equation with {Z t } WN(0,σ 2 ) and that the polynomials φ( ) and θ( ) have no common zeroes. Then {X t } is causal if and only if φ(z) 0 for all z C such that z 1. The coefficients {ψ j } j=0 in (2.1.19) are determined by the relation

20 13 ψ(z) = ψ j z j = θ(z)/φ(z), z 1. (2.1.21) j=0 Proof. See Brockwell and Davis (1991). Theorem Suppose that {X t } satisfies the ARMA(p,q) difference equation with {Z t } WN(0,σ 2 ) and that φ( ) and θ( ) have no common zeroes. Then {X t } is invertible if and only if θ(z) 0 for all z C such that z 1. The coefficients {π j } j=0 in (2.1.20) are determined by the relation π(z) = π j z j = φ(z)/θ(z), z 1. (2.1.22) j=0 Proof. See Brockwell and Davis (1991). 2.2 Distribution Classes of Discrete Random Variables This subsection overviews stochastic orderings (shapes) of discrete random variables. Kijima (1997) and Muller and Stoyan (2002) are comprehensive references on this topic. Let X be a discrete random variable taking values in {1, 2, 3,...} with probability vector a = (a i ) i=1, where a i = P[X = i] and i=1 a i = 1. If the support of X is finite, i.e., there is some N such that N i=1 a i = 1, then a i = 0 for all i > N. The hazard rate function of X is defined by h i = a i A i ; (2.2.1) where A i = P[X i] = k=i a k whenever A i > 0. Note that A i = 0 implies A j = 0 for all j > i. An interpretation of the hazard rate function is merely h i = P[X = i] P[X i] = P[X = i X i]; (2.2.2)

21 14 that is, h i is the probability that, conditional on survival up to time i, the lifetime X is equal to i. Definition (Increasing Hazard Rate). A discrete random variable X with distribution a = (a i ) i=1 is said to have an increasing hazard rate, denoted by X IHR or a IHR, if A i+1 2 Ai A i+2, i 0. (2.2.3) If the inequality is reversed, X is said to have decreasing hazard rate and is denoted by X DHR or a DHR. We note that, since A i = 0 implies that A j = 0 for j > i, X IHR if and only if with the convention 0/0 = 0. Since A i+1 A i A i+2 A i+1, i 0, (2.2.4) A i+1 A i = 1 a i A i = 1 h i, (2.2.5) provided that A i > 0, X is IHR if and only if the hazard rate function of X is increasing on the support of X. Similarly, it is easily seen that X DHR if and only if A i+1 /A i is increasing in i; that is, X is DHR if and only if the hazard rate function of X is decreasing. These observations justify the terms IHR and DHR in Definition Next, the likelihood ratio function of X is defined by whenever a i > 0. l i = a i+1 a i, (2.2.6) Definition (Decreasing Likelihood Ratio). A discrete random variable X with distribution a = (a i ) i=1 is said to have a decreasing likelihood ratio, denoted by X DLR or a DLR, if

22 15 a 2 i+1 a i a i+2. (2.2.7) If the inequality is reversed, X is said to have an increasing likelihood ratio and is denoted by X ILR or a ILR. Now suppose for simplicity that 0 < a 1 < 1. Then, if X DLR, a i = 0 implies that a j = 0 for all j > i. Hence X DLR implies that a i+1 a i = l i l i+1 = a i+2 a i+1, i 0, (2.2.8) with the convention 0/0 = 0; that is, the likelihood ratio function of X is decreasing. Similarly, if X ILR then a i > 0 for all i 0 and the likelihood ratio function of X is increasing. Definition (New Better than Used). A discrete random variable X with distribution a = (a i ) i=1 is called new better than used and is denoted by X NBU or a NBU, if A i+j A i A j, i,j 0. (2.2.9) If the inequality is reversed, X is called new worse than used and is denoted by X NWU or a NWU. The NBU property is written as P[X i] = A i A i+j A j = P[X i + j X j], j = 1, 2,..., (2.2.10) with the convention 0/0 = 0. The left-hand side in the above inequality denotes the survival probability of a new unit after i time intervals, whereas the right-hand side is the survival probability of an old unit with age j for i additional time periods. The term new better than used is thus interpreted as saying that a new item lasts stochastically longer than a used item. Analogous to (2.2.10), the NWU property is written as

23 16 P[X i] = A i A i+j A j = P[X i + j X j], j = 1, 2,..., (2.2.11) with the convention 0/0 = 0. The term new worse than used is thus interpreted as a new item lasts stochastically shorter than a used item. Theorem In the distribution classes of discrete random variables, we have (i) DLR IHR NBU; (ii) ILR DHR NWU. Proof. See Kijima (1997). The containments in Theorem are all strict. In particular, there are IHR distributions that are not DLR. 2.3 Monotone Markov Chains Markov chains are discrete-time stochastic processes that obey the so called Markov forgetfulness property in that the distribution of future states depends only on the current state and not on previous history of the process. The Markov property was proposed by A.A. Markov ( ) as part of his work on generalizing the classical limit theorems of probability. Markov chains have many applications in, for example, operations research, biology, engineering, and economics. We take the case of a discrete state space on the states i 0,i 1,i 2,.... Definition (Markov Property). The sequence {X n } n=0 is called a Markov chain on the states i 0,i 1,i 2,..., if for each n 0 and every i 0,...,i n and j 0, P[X n+1 = j X 0 = i 0,...,X n = i n ] = P[X n+1 = j X n = i n ] (2.3.1)

24 17 Given the whole history {X 0 = i 0,...,X n = i n }, the Markov property (2.3.1) states that the current state X n = i n determines the distributions of all future chain values X n+h for h 0. Definition (Monotone Markov chain). A Markov chain is said to be stochastically monotone if, when {X n } n=0 and {X n} n=0 are two realizations of the chain with X 0 X 0 stochastically (ie. P[X 0 > x] P[X 0 > x], for all x), then X n X n for all n 1 stochastically. A monotone Markov chain merely stipulates that a higher initial level produces a higher level at all other times. Example Suppose that {X n } n=0 is a stochastically monotone Markov chain on the states {0, 1, 2,...}. Then the return time to state 0, is NWU. Proof. τ 0 = inf n 1 {X n = 0 X 0 = 0} (2.3.2) P[τ 0 i + j X 0 = 0] = P[τ 0 i + j τ 0 i X 0 = 0] P[τ 0 i X 0 = 0] P[τ 0 i X 0 = 0] = P[τ 0 i + j τ 0 i X 0 = 0]P[τ 0 i X 0 = 0] P[τ 0 j X 0 = 0]P[τ 0 i X 0 = 0], (2.3.3) where the last line follows from stochastic monotonicity. Berenhaut and Lund (2002) discuss how to obtain a total variation Markov chain convergence rate from a chain with a DHR state, and give the following explicit rate and first constant. Theorem Suppose that {X n } n=0 is an ergodic (aperiodic, positive recurrent, irreducible) Markov chain on a countable state space and that the chain has a state k such that τ k (starting from X 0 = k) is DHR. Then for all n 0,

25 sup A ( P[X n A X 0 = k] π(a) n + 1 R F 1 18 ) R n F (2.3.4) where π is the unique stationary measure of the chain and R F denotes the radius of convergence of E[r τ k X0 = k]. Proof. See Berenhaut and Lund (2002). Computing R F in practice may be difficult. However, one can show that the same bound in Theorem applies to any r > 1 satisfying E[r τ k X0 = k] < : sup P[X n A X 0 = k] π(a) A ( ) n + 1 r n (2.3.5) r 1 Foster-Lyapunov drift methods (cf. Meyn and Tweedie, 1993), for example, allow practical identification of r > 1 such that E k [r τ k ] < with minimal effort. Stochastic shapes and Markov chains have been linked for over twenty years now (Keilson and Kester, 1978); however, connecting shapes with explicit convergence rates remains relatively unexplored, with the dissertation by Berenhaut (2000) being an exception. 2.4 Renewal Theory This section overviews some results in renewal theory and its applications. A stochastic process {N(t),t 0} is said to be a counting process if N(t) is the total number of events that have occurred up to time t. Then a counting process for which the times between successive events are independent and identically distributed with an arbitrary distribution is called a renewal process. For example, a Poisson process is a counting process for which the times between successive events are independent and identically distributed exponential random variables.

26 Definition If the sequence of nonnegative random variables {X 1,X 2,...} is independent and identically distributed, then the counting process {N(t), t 0} is said to be a renewal process. We will work in discrete time in what follows and take X i to be supported on {1, 2, 3,...}. We call the X i in this case lifetimes. For a renewal process having lifetimes {X i } i=1 with distribution given by P[X i = k] = f k, set S 0 = 0, and S n = n i=1 X i, for n 1. Then S n denotes the time of the nth renewal. The renewal sequence {u n } n=0 is defined by u n = k=0 P[S k = n], for n 0. Hence u n is the probability of a renewal occurring at time n. The convention u 0 = 1 is made. We assume that X 1 is not supported on a sublattice of {1, 2,...}, that the mean recurrent time µ def = E[X 1 ] <, and that X 1 is not degenerate in the sense that 19 The classical recurrent event relation is P[X 1 > 1] > 0. (2.4.1) n u n = f k u n k, n 1. (2.4.2) k=1 It can be shown that (2.4.2) defines a one-to-one map between probability mass functions on the positive integers and renewal probabilities. The basic renewal theorem (Erdös, Feller and Pollard, 1949) states that as n for any non-lattice lifetime X 1. 1 def u n E[X 1 ] = u (2.4.3) The conventional approach to obtain rates of convergence in (2.4.3) places restrictions on the tails of the sequence {q n } n=0, where q n = P[X 1 > n], usually in the form of moment conditions. For example, Feller (1957) proved that if E[X 2 1] <, then

27 20 ( ) def 1 e n = u n u = o n (2.4.4) as n. Theorem (Kendall 1959) An s > 1 exists such that E[s X 1 ] < if and only if an r > 1 exists satisfying e n = o(r n ). (2.4.5) Lindvall (1979) gives a coupling proof of Theorem However, the rate r in the statement of Theorem is not explicit. For instance, if E[s X 1 ] <, (2.4.5) may not hold with r = s. Our goal here is to investigate conditions under which explicit r > 1 can be identified satisfying (2.4.5), and to identify the largest such r if possible. It will be shown in Chapter 4, that under decreasing hazard rate lifetimes, a good convergence rate bound can be obtained. Berenhaut and Lund (2001) rehashed Heathcote (1967) and Malyshev and Spieksma (1995) to show that the optimal geometric convergence rate in (2.4.5) involves roots of a power series. For complex z with z < 1, define U(z) = u n z n and F(z) = P[X 1 = n]z n = E[z X 1 ]. (2.4.6) n=0 n=1 The relationship U(z) = (1 F(z)) 1 follows from the classical recurrent event equation (2.4.2) and relates the two power series in (2.4.6). The radius of convergence of the power series (z) = (u n u )z n (2.4.7) n=0

28 21 is denoted by R and is the optimal geometric decay rate of u n to u. R exceeds unity by Kendall s renewal theorem whenever X 1 has a finite geometric moment. A useful identity (Heathcote, 1967) is (z) = F (1) (z) 1 F(z) 1 = ( E[X 2 1 ] E[X 1 ] 2E[X 1 ] 2 ) F (2) (z) F (1) (z) (2.4.8) for z < 1, where F (1) (z) = E[z X(1) 1 ] and F (2) (z) = E[z X(2) 1 ] are the generating functions of the first two distributions derived from the tails of X 1 : P[X (1) 1 = n] = E[X 1 ] 1 P[X 1 > n] for n 0 and P[X (2) = n] = E[X (1) 1 ] 1 = P[τ (1) 1 > n] for n 0. One can also obtain F (1) (z) = F(z) 1 E[τ 1 ](z 1) and F (2) (z) = F (1) (z) 1 E[τ (1) 1 ](z 1). (2.4.9) The radius of convergence of F, denoted by R F > 1, is also the radius of convergence of F (1) and F (2). The decay rate R can be related to the location of zeros (possibly complex) of F (1). Since F (1) is also a probability generating function, F (1) (1) = 1 and z = 1 is not a zero of F (1). Let r 0 = z 0 denote the magnitude of the smallest nonzero root of F (1) (z 0 is a solution to F (1) (z 0 ) = 0) and suppose that r 0 < R F <. We now argue that R = r 0. First, note that F (1) and F (2) cannot have a common zero inside their common radius of convergence. For if F (1) (z 0 ) = 0 for some z 0 < R F, then (2.4.9) gives F (2) (z 0 ) = E[τ (1) 1 ] 1 0. Hence, F (2) (z 0 )/F (1) (z 0 ) =, and because power series that agree on the disk z < 1 must agree elsewhere (uniqueness of Taylor expansions), (2.4.8) gives (z 0 ) =. A triangle inequality establishes the absolute divergence = (z 0 ) u n u r 0 n. (2.4.10) n=0

29 22 Hence, R r 0. To argue that R r 0, return to (2.4.8) and use the fact that ratios of analytic functions are analytic in regions where the denominator has no zeros. Hence, (z) is analytic in z < r 0 and R r 0. Many applications of renewal theory are made to Markov chains. Here X 1 typically represents the first return time to a fixed state starting from that same state in the chain. It will be shown that good convergence rates for reversible Markov chains can be obtained from distributional orderings of X 1. Renewal theory and stochastic orderings have been previously considered by Kaluza (1928), Shanhbag (1977), Brown (1980, 1981), Shanthikumar (1988), Hansen and Frenk (1991) and Sengupta et al. (1995). In general, these authors explore monotonicities and inequalities for the renewal function, not rates of convergence. Liggett (1989) and Embrechts and Omey (1984) do obtain some limited results on geometric renewal convergence rates. 2.5 References [1] Berenhaut, K. S. (2000). Geometric Renewal Convergence Rates and Discrete Lifetime Distribution Classes, Ph.D. Dissertation, University of Georgia. [2] Berenhaut, K. S., and Lund, R. B. (2001). Geometric renewal convergence rates from hazard rates, Journal of Applied Probability, 38, [3] Berenhaut, K. S., and Lund, R. B. (2002). Renewal convergence rates for DHR and NWU lifetimes, Probability in the Engineering and Informational Sciences, 16, [4] Brockwell, P. J. and Davis, R. A. (1991). Time Series: Theory and Methods, second edition, New York: Springer.

30 [5] Brown, M. (1980). Bounds, inequalities, and monotonicity properties for some specialized renewal processes, The Annals of Probability, 8, [6] Brown, M. (1981). Further monotonicity properties for specialized renewal processes, The Annals of Probability, 9, [7] Embrechts, P., and Omey, E. (1984). Functions of power series, Yokohama Mathematical Journal, 32, [8] Erdös, P., Feller, W., and Pollard, H. (1945). A property of power series with positive coefficients, Bulletin of the American Mathematical Society, 55, [9] Feller W. (1957). An Introduction to Probability Theory and Its Applications, second edition, New York: John Wiley & Sons. [10] Hansen, B. G., and Frenk, J. B. G. (1991). Some monotonicity properties of the delayed renewal function, Journal of Applied Probability, 28, [11] Heathcote, C. R. (1967). Complete exponential convergence and related topics, Journal of Applied Probability, 4, [12] Kaluza, T. (1928). Uber die Koeffizienten Reziproks Potenzreihen, Mathematische Zeitschrift, 28, Keilson, J., and A. Kester (1978). Unimodality preservation in Markov chains, Stochastic Processes and their Applications, 7, [13] Kendall, D. G. (1959). Unitary dilations of Markov transition operators and the corresponding integral representations for transition probability matrices,

31 In: Probability and Statistics, Edited by U. Grenander, , New York: Wiley. 24 [14] Kijima, M. (1997). Markov Processes for Stochastic Modeling, London: Chapman and Hall. [15] Liggett, T. (1989). Total positivity and renewal theory, In: Probability, Statistics, and Mathematics, Edited by T. W. Anderson, K. B. Athreya, and D. L. Iglehart, , Boston: Academic Press. [16] Lindvall, T. (1979). On coupling of discrete renewal processes, Z. Wahrsch. Verw. Gebiete, 48, [17] Malyshev, V. A., and Spieksma, F. M. (1995). Intrinsic convergence rate of countable Markov chains, Markov Processes and Related Fields,1, [18] Marshall, A. W., and Shaked, M. (1986). NBU processes with general state space. Math. Operat. Res. 11, [19] Meyn, S. P. and Tweedie, R. L. (1993). Markov Chains and Stochastic Stability, New York: Springer-Verlag. [20] Muller, A., and Stoyan, D. (2002). Comparison Methods for Stochastic Models and Risks, New York: Wiley. [21] Ramsey, F. L. (1974). Characterization of the partial autocorrelation function, Ann. of Statist., 2,

32 25 [22] Sengupta, D., Chatterjee, A., and Chakraborty, B. (1995). Reliability bounds and other inequalities for discrete life distributions, Microelectronics and Reliability, 35, [23] Shanhbag, D. N. (1977). On renewal sequences, Bulletin of the London Mathematical Society, 9, [24] Shanthikumar, J. G. (1988). DFR property of first-passage times and its preservation under geometric compounding, Annals of Probability, 16,

33 Chapter 3 Shape Orderings for Stationary Time Series 1 1 Y. Zhao and R. Lund Submitted to Journal of Applied Probability, 10/

34 27 Abstract This paper introduces shape orderings for stationary time series autocorrelation and partial autocorrelation functions and explores some of their convergence rate ramifications. The shapes explored include increasing likelihood ratio, decreasing hazard rate, and new better than used, structures familiar from stochastic processes and reliability settings. Examples of autoregressive moving-average time series having these shapes are first presented. The shapes are then applied to obtain explicit geometric convergence rates of several forecasting quantities. Key Words and Phrases: Autocorrelation; Convergence Rate; Decreasing Hazard Rate; Increasing Likelihood Ratio; Mean Squared Error; New Better Than Used; Partial Autocorrelation. 3.1 Introduction This paper introduces stochastic shape orderings for the autocorrelation (ACF) and partial autocorrelation (PACF) functions of stationary time series and studies some of their convergence rate ramifications. The orderings introduced are familiar from stochastic process and reliability settings and include new worse than used, new better than used, increasing hazard rate, decreasing hazard rate, increasing likelihood ratio, and decreasing likelihood ratio. The utility of such orderings can be appreciated by browsing Keilson and Kester (1978), Brown (1980), Shaked and Shanthikumar (1994), Liggett (1989), Hansen and Frenk (1991), Sengupta et al. (1995), Kijima (1997), Berenhaut and Lund (2001), and Muller and Stoyan (2003). The above list of ordering types is not exhaustive: decreasing reversed hazard rate, new better than used in expectation, increasing hazard rate average, etc. orderings have also proven useful in applications (cf. Shaked and Shanthikumar 1987, 1994; Kijima 1997; Muller and Stoyan 2003).

35 28 The main purpose of this article is to introduce stochastic orderings into the time series analyst s toolbox. We are unaware of any previous literature with this slant. Given the utility of ordering methods in stochastic processes, the theme is expected to prove fruitful. Nonetheless, we do not make any attempt at completing the issue here; indeed, the applications pursued here reside only with explicit convergence rate bounds for mean squared prediction errors and Innovations Algorithm coefficients. It is expected that shape orderings will eventually be useful in likelihood computations, forecasting, and quantifying how close weighted and least squares estimators are. The remainder of this paper proceeds as follows. Section 2 clarifies the orderings that we study. Section 3 gives examples of common autoregressive moving-average (ARMA(p, q)) time series models which obey one or more of the introduced orderings. Section 4 applies the shape ideas by deriving explicit geometric convergence rates for several quantities encountered in forecasting. 3.2 Definition of Orderings Let {X t } be a zero mean covariance stationary series with autocovariance γ(h) = Cov(X t+h,x t ) at lag h. We denote the lag h autocorrelations and partial autocorrelations by ρ(h) = Corr(X t+h,x t ) and α(h) = Corr(X h+1,x 1 X 2,X 3,...,X h ), respectively. Clarifying, a conditional correlation refers to correlation after adjustment for best linear prediction given X 2,...,X h : ( ) α(h) = Corr X h+1 P sp{x2,...,x h }(X h+1 ),X 1 P sp{x2,...,x h }(X 1 ), (3.2.1) where sp( ) denotes closed linear span and P( ) indicates a linear prediction.

36 The autocorrelation function ρ( ) is said to be new better than used (NBU) if 29 If the inequality in (3.2.2) is reversed to ρ(i + j) ρ(i)ρ(j), i,j 0. (3.2.2) ρ(i + j) ρ(i)ρ(j), i,j 0, then {X t } is said to have a new worse than used (NWU) ACF. The series {X t } is said to have a decreasing likelihood ratio (DLR) ACF if If the inequality in (3.2.3) is reversed to ρ(h + 1) 2 ρ(h)ρ(h + 2), h 0. (3.2.3) ρ(h + 1) 2 ρ(h)ρ(h + 2), h 0, (3.2.4) then the ACF of {X t } is said to have an increasing likelihood ratio (ILR). The series {X t } is said to have an increasing hazard rate (IHR) ACF if the hazard rates h i defined by h i = k=i ρ(i) i 0, (3.2.5) ρ(k), are nondecreasing in i; if (3.2.5) is decreasing as i increases, then {X t } is said to have a decreasing hazard rate (DHR) ACF. We say that {X t } has a monotone autocorrelation function if ρ(h) is decreasing as h increases; {X t } is said to have a convex autocorrelation function if ρ(h) is convex in h in the sense that ρ(h) 2ρ(h 1) + ρ(h 2) 0, h 2.

37 30 Since many convex sequences with ρ(0) > 0 are non-negative definite (Polya s criterion), such sequences are indeed legitimate stationary autocovariance functions (cf. Problem 26.3 in Billingsley, 1995, for sufficient conditions). The above definitions are analogues of orderings taking the same name for a univariate random variable X {0, 1, 2...}: we have merely interchanged P(X > h) with ρ(h). Some properties follow immediately from stochastic processes theory. For example, any DHR ACF is also NWU; however, there exist NWU ACFs that are not DHR (cf. Kijima 1997; Berenhaut and Lund 2001). Likewise, any IHR ACF is NBU, but such containment is not exclusive. As ρ(h) = γ(0) 1 γ(h), the above shapes are also meaningful for the ACVF without modification. For shapes of the PACF, we merely exchange ρ( ) with α( ). For example, {X t } is said to have an NBU PACF if α(i + j) α(i)α(j), i,j Examples This section presents examples of ARMA(p, q) time series models possessing some of the above orderings. As a very simple first example, consider a first-order causal autoregression {X t } satisfying X t = φx t 1 + Z t, where {Z t } is zero mean white noise with variance σ 2. Here, ρ(h) = φ h for h 0 ( φ < 1 is due to causality). Hence, the AR(1) ACF is NWU, NBU, ILR, and DLR. When φ (0, 1), the AR(1) ACF is monotone and convex. It is not too hard to show that a first order autoregression is the only stationary series with an ACF that is both NWU and NBU.

38 31 Example Every first-order moving-average (MA(1)) has a DLR ACF. Proof. Write the MA(1) as X t = Z t + θz t 1, (3.3.1) where {Z t } is zero mean white noise with variance σ 2. The ACF of {X t } is ρ(h) = 1 {0} (h)+θ/(1+θ 2 )1 {±1} (h). Hence, ρ 2 (0) = 1 and ρ(1) 2 = θ 2 /[1+θ 2 ] 2. As ρ(h)ρ(h+ 2) = θ 2 /[1 + θ 2 ] 2 for h = 1 and zero otherwise, the claim now follows. Example Every second-order causal autoregression (AR(2)) has a DLR ACVF. Proof. Here, {X t } is a solution to the difference equation X t φ 1 X t 1 φ 2 X t 2 = Z t, (3.3.2) where {Z t } is zero mean white noise with variance σ 2. Factoring the AR(2) polynomial into its roots gives (1 ξ 1 1 B)(1 ξ 1 2 B)X t = Z t, where B is the usual backshift operator and causality implies that ξ 1 > 1 and ξ 2 > 1. We proceed only with the case of unequal roots: ξ 1 ξ 2. Using the relations φ 1 = ξ ξ 1 2 and φ 2 = ξ 1 1 ξ 1 2 in a difference equationbased expression for AR(2) autocovariances (cf. Brockwell and Davis 1991, Chapter 3) allows us to write γ(h) = σ 2 ξ 2 1ξ 2 2 (ξ 1 ξ 2 1)(ξ 2 ξ 1 ) [(ξ2 1 1) 1 ξ 1 h 1 (ξ 2 2 1) 1 ξ 1 h 2 ]. (3.3.3) Writing ξ 1 = re iθ and ξ 2 = re iθ for some θ (0,π] and r > 0 (ξ 1 and ξ 2 are complex conjugates) in (3.3.3) gives

39 32 γ(h) = σ 2 r 4 r h sin(hθ + ψ) (r 2 1)(r 4 2r 2 cos(2θ) + 1) 1 2 sin θ, where tan(ψ) = (r 2 + 1)(r 2 1) 1 tan(θ) and cos(ψ) has the same sign as cos(θ). Hence, and ρ(h) = r h sin(hθ + ψ) sin ψ ( ) sin(hθ) = tanψ + cos(hθ) r h (3.3.4) r i+j ρ(i)ρ(j) = = ( ) sin(iθ) sin(jθ) sin(iθ) cos(jθ) + cos(iθ) sin(jθ) + + cos(iθ) cos(jθ) tan 2 ψ tanψ [ ( sin((i + j)θ) 1 + tanψ 2 tan 2 ψ + 1 ) cos((i j)θ) + 2 ( ) ] cos((i + j)θ). (3.3.5) 2 tan 2 ψ Equation (3.3.5) gives the identities r 2(i+1) ρ(i)ρ(i + 2) = + [ ( sin(2(i + 1)θ) 1 + tanψ 2 tan 2 ψ + 1 ) cos(2θ) 2 ( ) ] cos(2(i + 1)θ) 2 tan 2 ψ (3.3.6) and

40 33 r 2(i+1) ρ 2 (i + 1) = = = = [ ] 2 sin((i + 1)θ) + cos((i + 1)θ) tanψ [ ] sin 2 ((i + 1)θ) + cos 2 2 sin((i + 1)θ) cos((i + 1)θ) ((i + 1)θ) + tan 2 θ tanψ [ ] 1 cos(2(i + 1)θ) 1 + cos(2(i + 1)θ) sin(2(i + 1)θ) tan 2 ψ 2 tanψ [ ( ) sin(2(i + 1)θ) 1 + tanψ 2 1 cos(2(i + 1)θ) + 2 tan 2 ψ ( )] (3.3.7) 2 tan 2 ψ Combining (3.3.6) and (3.3.7) gives ρ 2 (i + 1) ρ(i)ρ(i + 2) = ( ) (1 cos(2θ))r 2(i+1). 2 tan 2 ψ As 1 cos(2θ) 0, the DLR property of causal AR(2) ACFs now follows. Example A causal and invertible ARMA(1,1) series has a PACF whose square is monotonically decreasing and ILR. Proof. The ARMA(1,1) difference equation is X t φx t 1 = Z t + θz t 1, (3.3.8) where {Z t } is zero mean white noise with variance σ 2. Causality and invertibility imply that φ < 1 and θ < 1. The partial autocorrelation function of an ARMA(1,1) series can be explicitly identified from the result of Problem 5.13 in Brockwell and Davis (1991): α 2 (n) = The claimed monotonicity follows from (3.3.9) since θ 2n 2 (1 θ 2 ) 2 (θ + φ) 2 (1 + θφ) 2 [(θ + φ) 2 (1 θ 2n ) + (1 φ 2 )(1 θ 2 )] 2. (3.3.9)

41 34 α 2 (n 1) α 2 (n) = 1 θ 2 [(θ + φ) 2 (1 θ 2n ) + (1 φ 2 )(1 θ 2 )] 2 [(θ + φ) 2 (1 θ 2n 2 ) + (1 φ 2 )(1 θ 2 )] 2 1 θ 2 1, (3.3.10) which implies that α 2 (n) is nonincreasing in n. From this monotonicity, the ILR claim can be verified to follow. 3.4 Convergence Rates This section explores convergence rate consequences of the above orderings in onestep-ahead linear prediction settings. This topic is also studied in Section of Pourahamadi (2001). For notation, let ˆX t+1 = P(X t+1 X 1,...,X t ) = P sp{x1,...,x t}(x t+1 ) (3.4.1) be the one-step-ahead linear prediction of X t+1 from elements in the span of X 1,...,X t and let v t = E[(X t+1 ˆX t+1 ) 2 ] denote its unconditional mean squared prediction error. Let v = lim t v t denote the limiting mean squared prediction error; this limit exists as v t is nonincreasing in t. The mean squared prediction error can be expressed via the square of the PACF: t v t = γ(0) (1 α 2 (j)) (3.4.2) j=1

42 35 (cf. Proposition in Brockwell and Davis 1991 and Section 7.5 of Pourahamadi 2001). To avoid trite work, we assume that the covariance matrix of (X 1,...,X t ) is invertible for all positive integers t. Sufficient for this is merely that γ(h) 0 as h (cf. Proposition of Brockwell and Davis, 1991), as is the case for any causal ARMA(p,q) series. From (3.4.2), we deduce that v t v = γ(0) γ(0) [ t (1 α 2 (j)) 1 j=1 t (1 α 2 (j)) j=1 j=t+1 j=t+1 (1 α 2 (j)) α 2 (j), (3.4.3) where the inequality 1 j=t+1 (1 α2 (j)) j=t+1 α2 (j) has been applied. Note that (3.4.3) is tight for an autoregression of order p; specifically, the righthand side of (3.4.3) is zero for lags t > p (implying that v t σ 2 = v for t > p). Now suppose that a shape structure is imposed; specifically, consider the case where {X t } has DLR PACF. Then α(t + h) α(t)α h (1) for all t,h 0. Using this in (3.4.3) produces a very clean and explicit convergence bound: ] [ t ] v t v γ(0)α 2 (1) (1 α 2 (j)) α 2 (t) j=2 γ(0)α 2 (t), (3.4.4) where the last line follows after 0 t j=2 (1 α2 (j)) 1 and α 2 (1) 1 are applied. It is important to note that (3.4.4) holds for each and every t and as such is a convergence rate bound. This differs radically in structure from an asymptotic approximation that is applicable for large n only. One could pursue other shapes and inequalities, and we do this to an extent in the MA(1) and ARMA(1,1) examples below; however, the important theme is simply

43 36 that a shape constraint can provide clean and explicit convergence rates for the mean squared prediction errors. It is worth commenting that the shapes studied here are also useful in obtaining explicit geometric convergence rates of Markov chains to stationarity (cf. Lund, Meyn, and Tweedie 1996; Berenhaut and Lund 2001). Now if {X t } is an ARMA(p,q) series satisfying the difference equation X t φ 1 X t 1... φ p X t p = Z t + θ 1 Z t θ q Z t q, where {Z t } is zero mean white noise with variance σ 2 and the AR and MA polynomials are causal and invertible, then the classic one-step-ahead recursive prediction formula is ˆX n+1 = p q φ k X n+1 k + θ n,k ( ˆX n+1 k X n+1 k ), n max(p,q). (3.4.5) k=1 k=1 The θ n,k s are identified by applying the Innovations Algorithm to a strategic linear transform of {X t } (cf. Ansley 1979; Brockwell and Davis 1991, Chapter 5): θ n,k = v 1 n k E[X n(x n+1 k ˆX n+1 k )], 1 k q; n max(p,q). (3.4.6) In the case of an MA(1) series, one has an ILR PACF structure. This is harder to work with than DLR structures, but can be handled as we show in the next example. Example The θ n,1 in an invertible MA(1) series converge to θ with explicit convergence bound rate Proof. Equation (3.4.6) gives [ θ n,1 θ ] θ n. (3.4.7) 1 θ 2

STAT 443 Final Exam Review. 1 Basic Definitions. 2 Statistical Tests. L A TEXer: W. Kong

STAT 443 Final Exam Review. 1 Basic Definitions. 2 Statistical Tests. L A TEXer: W. Kong STAT 443 Final Exam Review L A TEXer: W Kong 1 Basic Definitions Definition 11 The time series {X t } with E[X 2 t ] < is said to be weakly stationary if: 1 µ X (t) = E[X t ] is independent of t 2 γ X

More information

3. ARMA Modeling. Now: Important class of stationary processes

3. ARMA Modeling. Now: Important class of stationary processes 3. ARMA Modeling Now: Important class of stationary processes Definition 3.1: (ARMA(p, q) process) Let {ɛ t } t Z WN(0, σ 2 ) be a white noise process. The process {X t } t Z is called AutoRegressive-Moving-Average

More information

STAT STOCHASTIC PROCESSES. Contents

STAT STOCHASTIC PROCESSES. Contents STAT 3911 - STOCHASTIC PROCESSES ANDREW TULLOCH Contents 1. Stochastic Processes 2 2. Classification of states 2 3. Limit theorems for Markov chains 4 4. First step analysis 5 5. Branching processes 5

More information

Applied time-series analysis

Applied time-series analysis Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna October 18, 2011 Outline Introduction and overview Econometric Time-Series Analysis In principle,

More information

γ 0 = Var(X i ) = Var(φ 1 X i 1 +W i ) = φ 2 1γ 0 +σ 2, which implies that we must have φ 1 < 1, and γ 0 = σ2 . 1 φ 2 1 We may also calculate for j 1

γ 0 = Var(X i ) = Var(φ 1 X i 1 +W i ) = φ 2 1γ 0 +σ 2, which implies that we must have φ 1 < 1, and γ 0 = σ2 . 1 φ 2 1 We may also calculate for j 1 4.2 Autoregressive (AR) Moving average models are causal linear processes by definition. There is another class of models, based on a recursive formulation similar to the exponentially weighted moving

More information

{ } Stochastic processes. Models for time series. Specification of a process. Specification of a process. , X t3. ,...X tn }

{ } Stochastic processes. Models for time series. Specification of a process. Specification of a process. , X t3. ,...X tn } Stochastic processes Time series are an example of a stochastic or random process Models for time series A stochastic process is 'a statistical phenomenon that evolves in time according to probabilistic

More information

Long-range dependence

Long-range dependence Long-range dependence Kechagias Stefanos University of North Carolina at Chapel Hill May 23, 2013 Kechagias Stefanos (UNC) Long-range dependence May 23, 2013 1 / 45 Outline 1 Introduction to time series

More information

Nonlinear time series

Nonlinear time series Based on the book by Fan/Yao: Nonlinear Time Series Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna October 27, 2009 Outline Characteristics of

More information

1 Linear Difference Equations

1 Linear Difference Equations ARMA Handout Jialin Yu 1 Linear Difference Equations First order systems Let {ε t } t=1 denote an input sequence and {y t} t=1 sequence generated by denote an output y t = φy t 1 + ε t t = 1, 2,... with

More information

Exponential decay rate of partial autocorrelation coefficients of ARMA and short-memory processes

Exponential decay rate of partial autocorrelation coefficients of ARMA and short-memory processes Exponential decay rate of partial autocorrelation coefficients of ARMA and short-memory processes arxiv:1511.07091v2 [math.st] 4 Jan 2016 Akimichi Takemura January, 2016 Abstract We present a short proof

More information

Part III Example Sheet 1 - Solutions YC/Lent 2015 Comments and corrections should be ed to

Part III Example Sheet 1 - Solutions YC/Lent 2015 Comments and corrections should be  ed to TIME SERIES Part III Example Sheet 1 - Solutions YC/Lent 2015 Comments and corrections should be emailed to Y.Chen@statslab.cam.ac.uk. 1. Let {X t } be a weakly stationary process with mean zero and let

More information

Time Series 2. Robert Almgren. Sept. 21, 2009

Time Series 2. Robert Almgren. Sept. 21, 2009 Time Series 2 Robert Almgren Sept. 21, 2009 This week we will talk about linear time series models: AR, MA, ARMA, ARIMA, etc. First we will talk about theory and after we will talk about fitting the models

More information

Covariances of ARMA Processes

Covariances of ARMA Processes Statistics 910, #10 1 Overview Covariances of ARMA Processes 1. Review ARMA models: causality and invertibility 2. AR covariance functions 3. MA and ARMA covariance functions 4. Partial autocorrelation

More information

Lecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications

Lecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications Lecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications Prof. Massimo Guidolin 20192 Financial Econometrics Winter/Spring 2018 Overview Moving average processes Autoregressive

More information

University of Oxford. Statistical Methods Autocorrelation. Identification and Estimation

University of Oxford. Statistical Methods Autocorrelation. Identification and Estimation University of Oxford Statistical Methods Autocorrelation Identification and Estimation Dr. Órlaith Burke Michaelmas Term, 2011 Department of Statistics, 1 South Parks Road, Oxford OX1 3TG Contents 1 Model

More information

The autocorrelation and autocovariance functions - helpful tools in the modelling problem

The autocorrelation and autocovariance functions - helpful tools in the modelling problem The autocorrelation and autocovariance functions - helpful tools in the modelling problem J. Nowicka-Zagrajek A. Wy lomańska Institute of Mathematics and Computer Science Wroc law University of Technology,

More information

Markov Chains. Arnoldo Frigessi Bernd Heidergott November 4, 2015

Markov Chains. Arnoldo Frigessi Bernd Heidergott November 4, 2015 Markov Chains Arnoldo Frigessi Bernd Heidergott November 4, 2015 1 Introduction Markov chains are stochastic models which play an important role in many applications in areas as diverse as biology, finance,

More information

Econometric Forecasting

Econometric Forecasting Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna October 1, 2014 Outline Introduction Model-free extrapolation Univariate time-series models Trend

More information

EASTERN MEDITERRANEAN UNIVERSITY ECON 604, FALL 2007 DEPARTMENT OF ECONOMICS MEHMET BALCILAR ARIMA MODELS: IDENTIFICATION

EASTERN MEDITERRANEAN UNIVERSITY ECON 604, FALL 2007 DEPARTMENT OF ECONOMICS MEHMET BALCILAR ARIMA MODELS: IDENTIFICATION ARIMA MODELS: IDENTIFICATION A. Autocorrelations and Partial Autocorrelations 1. Summary of What We Know So Far: a) Series y t is to be modeled by Box-Jenkins methods. The first step was to convert y t

More information

ECON 616: Lecture 1: Time Series Basics

ECON 616: Lecture 1: Time Series Basics ECON 616: Lecture 1: Time Series Basics ED HERBST August 30, 2017 References Overview: Chapters 1-3 from Hamilton (1994). Technical Details: Chapters 2-3 from Brockwell and Davis (1987). Intuition: Chapters

More information

Lecture 1: Brief Review on Stochastic Processes

Lecture 1: Brief Review on Stochastic Processes Lecture 1: Brief Review on Stochastic Processes A stochastic process is a collection of random variables {X t (s) : t T, s S}, where T is some index set and S is the common sample space of the random variables.

More information

Levinson Durbin Recursions: I

Levinson Durbin Recursions: I Levinson Durbin Recursions: I note: B&D and S&S say Durbin Levinson but Levinson Durbin is more commonly used (Levinson, 1947, and Durbin, 1960, are source articles sometimes just Levinson is used) recursions

More information

FE570 Financial Markets and Trading. Stevens Institute of Technology

FE570 Financial Markets and Trading. Stevens Institute of Technology FE570 Financial Markets and Trading Lecture 5. Linear Time Series Analysis and Its Applications (Ref. Joel Hasbrouck - Empirical Market Microstructure ) Steve Yang Stevens Institute of Technology 9/25/2012

More information

Lesson 9: Autoregressive-Moving Average (ARMA) models

Lesson 9: Autoregressive-Moving Average (ARMA) models Lesson 9: Autoregressive-Moving Average (ARMA) models Dipartimento di Ingegneria e Scienze dell Informazione e Matematica Università dell Aquila, umberto.triacca@ec.univaq.it Introduction We have seen

More information

MATH 5075: Time Series Analysis

MATH 5075: Time Series Analysis NAME: MATH 5075: Time Series Analysis Final For the entire test {Z t } WN(0, 1)!!! 1 1) Let {Y t, t Z} be a stationary time series with EY t = 0 and autocovariance function γ Y (h). Assume that a) Show

More information

For a stochastic process {Y t : t = 0, ±1, ±2, ±3, }, the mean function is defined by (2.2.1) ± 2..., γ t,

For a stochastic process {Y t : t = 0, ±1, ±2, ±3, }, the mean function is defined by (2.2.1) ± 2..., γ t, CHAPTER 2 FUNDAMENTAL CONCEPTS This chapter describes the fundamental concepts in the theory of time series models. In particular, we introduce the concepts of stochastic processes, mean and covariance

More information

Stochastic Processes: I. consider bowl of worms model for oscilloscope experiment:

Stochastic Processes: I. consider bowl of worms model for oscilloscope experiment: Stochastic Processes: I consider bowl of worms model for oscilloscope experiment: SAPAscope 2.0 / 0 1 RESET SAPA2e 22, 23 II 1 stochastic process is: Stochastic Processes: II informally: bowl + drawing

More information

3 Theory of stationary random processes

3 Theory of stationary random processes 3 Theory of stationary random processes 3.1 Linear filters and the General linear process A filter is a transformation of one random sequence {U t } into another, {Y t }. A linear filter is a transformation

More information

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley Time Series Models and Inference James L. Powell Department of Economics University of California, Berkeley Overview In contrast to the classical linear regression model, in which the components of the

More information

Monotonicity and Aging Properties of Random Sums

Monotonicity and Aging Properties of Random Sums Monotonicity and Aging Properties of Random Sums Jun Cai and Gordon E. Willmot Department of Statistics and Actuarial Science University of Waterloo Waterloo, Ontario Canada N2L 3G1 E-mail: jcai@uwaterloo.ca,

More information

Levinson Durbin Recursions: I

Levinson Durbin Recursions: I Levinson Durbin Recursions: I note: B&D and S&S say Durbin Levinson but Levinson Durbin is more commonly used (Levinson, 1947, and Durbin, 1960, are source articles sometimes just Levinson is used) recursions

More information

at least 50 and preferably 100 observations should be available to build a proper model

at least 50 and preferably 100 observations should be available to build a proper model III Box-Jenkins Methods 1. Pros and Cons of ARIMA Forecasting a) need for data at least 50 and preferably 100 observations should be available to build a proper model used most frequently for hourly or

More information

Discrete time processes

Discrete time processes Discrete time processes Predictions are difficult. Especially about the future Mark Twain. Florian Herzog 2013 Modeling observed data When we model observed (realized) data, we encounter usually the following

More information

TIME SERIES AND FORECASTING. Luca Gambetti UAB, Barcelona GSE Master in Macroeconomic Policy and Financial Markets

TIME SERIES AND FORECASTING. Luca Gambetti UAB, Barcelona GSE Master in Macroeconomic Policy and Financial Markets TIME SERIES AND FORECASTING Luca Gambetti UAB, Barcelona GSE 2014-2015 Master in Macroeconomic Policy and Financial Markets 1 Contacts Prof.: Luca Gambetti Office: B3-1130 Edifici B Office hours: email:

More information

6.3 Forecasting ARMA processes

6.3 Forecasting ARMA processes 6.3. FORECASTING ARMA PROCESSES 123 6.3 Forecasting ARMA processes The purpose of forecasting is to predict future values of a TS based on the data collected to the present. In this section we will discuss

More information

Lecture 1: Fundamental concepts in Time Series Analysis (part 2)

Lecture 1: Fundamental concepts in Time Series Analysis (part 2) Lecture 1: Fundamental concepts in Time Series Analysis (part 2) Florian Pelgrin University of Lausanne, École des HEC Department of mathematics (IMEA-Nice) Sept. 2011 - Jan. 2012 Florian Pelgrin (HEC)

More information

Gaussian processes. Basic Properties VAG002-

Gaussian processes. Basic Properties VAG002- Gaussian processes The class of Gaussian processes is one of the most widely used families of stochastic processes for modeling dependent data observed over time, or space, or time and space. The popularity

More information

Introduction to Stochastic processes

Introduction to Stochastic processes Università di Pavia Introduction to Stochastic processes Eduardo Rossi Stochastic Process Stochastic Process: A stochastic process is an ordered sequence of random variables defined on a probability space

More information

Econometría 2: Análisis de series de Tiempo

Econometría 2: Análisis de series de Tiempo Econometría 2: Análisis de series de Tiempo Karoll GOMEZ kgomezp@unal.edu.co http://karollgomez.wordpress.com Segundo semestre 2016 III. Stationary models 1 Purely random process 2 Random walk (non-stationary)

More information

Ch 4. Models For Stationary Time Series. Time Series Analysis

Ch 4. Models For Stationary Time Series. Time Series Analysis This chapter discusses the basic concept of a broad class of stationary parametric time series models the autoregressive moving average (ARMA) models. Let {Y t } denote the observed time series, and {e

More information

Contents. 1 Time Series Analysis Introduction Stationary Processes State Space Modesl Stationary Processes 8

Contents. 1 Time Series Analysis Introduction Stationary Processes State Space Modesl Stationary Processes 8 A N D R E W T U L L O C H T I M E S E R I E S A N D M O N T E C A R L O I N F E R E N C E T R I N I T Y C O L L E G E T H E U N I V E R S I T Y O F C A M B R I D G E Contents 1 Time Series Analysis 5

More information

Convergence Rates for Renewal Sequences

Convergence Rates for Renewal Sequences Convergence Rates for Renewal Sequences M. C. Spruill School of Mathematics Georgia Institute of Technology Atlanta, Ga. USA January 2002 ABSTRACT The precise rate of geometric convergence of nonhomogeneous

More information

STAD57 Time Series Analysis. Lecture 8

STAD57 Time Series Analysis. Lecture 8 STAD57 Time Series Analysis Lecture 8 1 ARMA Model Will be using ARMA models to describe times series dynamics: ( B) X ( B) W X X X W W W t 1 t1 p t p t 1 t1 q tq Model must be causal (i.e. stationary)

More information

A Functional Central Limit Theorem for an ARMA(p, q) Process with Markov Switching

A Functional Central Limit Theorem for an ARMA(p, q) Process with Markov Switching Communications for Statistical Applications and Methods 2013, Vol 20, No 4, 339 345 DOI: http://dxdoiorg/105351/csam2013204339 A Functional Central Limit Theorem for an ARMAp, q) Process with Markov Switching

More information

Difference equations. Definitions: A difference equation takes the general form. x t f x t 1,,x t m.

Difference equations. Definitions: A difference equation takes the general form. x t f x t 1,,x t m. Difference equations Definitions: A difference equation takes the general form x t fx t 1,x t 2, defining the current value of a variable x as a function of previously generated values. A finite order

More information

Stat 248 Lab 2: Stationarity, More EDA, Basic TS Models

Stat 248 Lab 2: Stationarity, More EDA, Basic TS Models Stat 248 Lab 2: Stationarity, More EDA, Basic TS Models Tessa L. Childers-Day February 8, 2013 1 Introduction Today s section will deal with topics such as: the mean function, the auto- and cross-covariance

More information

Introduction to ARMA and GARCH processes

Introduction to ARMA and GARCH processes Introduction to ARMA and GARCH processes Fulvio Corsi SNS Pisa 3 March 2010 Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 1 / 24 Stationarity Strict stationarity: (X 1,

More information

Autoregressive Moving Average (ARMA) Models and their Practical Applications

Autoregressive Moving Average (ARMA) Models and their Practical Applications Autoregressive Moving Average (ARMA) Models and their Practical Applications Massimo Guidolin February 2018 1 Essential Concepts in Time Series Analysis 1.1 Time Series and Their Properties Time series:

More information

Introduction to Time Series Analysis. Lecture 7.

Introduction to Time Series Analysis. Lecture 7. Last lecture: Introduction to Time Series Analysis. Lecture 7. Peter Bartlett 1. ARMA(p,q) models: stationarity, causality, invertibility 2. The linear process representation of ARMA processes: ψ. 3. Autocovariance

More information

Stochastic processes: basic notions

Stochastic processes: basic notions Stochastic processes: basic notions Jean-Marie Dufour McGill University First version: March 2002 Revised: September 2002, April 2004, September 2004, January 2005, July 2011, May 2016, July 2016 This

More information

On the Convolution Order with Reliability Applications

On the Convolution Order with Reliability Applications Applied Mathematical Sciences, Vol. 3, 2009, no. 16, 767-778 On the Convolution Order with Reliability Applications A. Alzaid and M. Kayid King Saud University, College of Science Dept. of Statistics and

More information

Forecasting. Simon Shaw 2005/06 Semester II

Forecasting. Simon Shaw 2005/06 Semester II Forecasting Simon Shaw s.c.shaw@maths.bath.ac.uk 2005/06 Semester II 1 Introduction A critical aspect of managing any business is planning for the future. events is called forecasting. Predicting future

More information

Covariance Stationary Time Series. Example: Independent White Noise (IWN(0,σ 2 )) Y t = ε t, ε t iid N(0,σ 2 )

Covariance Stationary Time Series. Example: Independent White Noise (IWN(0,σ 2 )) Y t = ε t, ε t iid N(0,σ 2 ) Covariance Stationary Time Series Stochastic Process: sequence of rv s ordered by time {Y t } {...,Y 1,Y 0,Y 1,...} Defn: {Y t } is covariance stationary if E[Y t ]μ for all t cov(y t,y t j )E[(Y t μ)(y

More information

Empirical Market Microstructure Analysis (EMMA)

Empirical Market Microstructure Analysis (EMMA) Empirical Market Microstructure Analysis (EMMA) Lecture 3: Statistical Building Blocks and Econometric Basics Prof. Dr. Michael Stein michael.stein@vwl.uni-freiburg.de Albert-Ludwigs-University of Freiburg

More information

Faithful couplings of Markov chains: now equals forever

Faithful couplings of Markov chains: now equals forever Faithful couplings of Markov chains: now equals forever by Jeffrey S. Rosenthal* Department of Statistics, University of Toronto, Toronto, Ontario, Canada M5S 1A1 Phone: (416) 978-4594; Internet: jeff@utstat.toronto.edu

More information

STAT 248: EDA & Stationarity Handout 3

STAT 248: EDA & Stationarity Handout 3 STAT 248: EDA & Stationarity Handout 3 GSI: Gido van de Ven September 17th, 2010 1 Introduction Today s section we will deal with the following topics: the mean function, the auto- and crosscovariance

More information

EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY

EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA, 011 MODULE 3 : Stochastic processes and time series Time allowed: Three Hours Candidates should answer FIVE questions. All questions carry

More information

Time Series 3. Robert Almgren. Sept. 28, 2009

Time Series 3. Robert Almgren. Sept. 28, 2009 Time Series 3 Robert Almgren Sept. 28, 2009 Last time we discussed two main categories of linear models, and their combination. Here w t denotes a white noise: a stationary process with E w t ) = 0, E

More information

Review Session: Econometrics - CLEFIN (20192)

Review Session: Econometrics - CLEFIN (20192) Review Session: Econometrics - CLEFIN (20192) Part II: Univariate time series analysis Daniele Bianchi March 20, 2013 Fundamentals Stationarity A time series is a sequence of random variables x t, t =

More information

Chapter 12: An introduction to Time Series Analysis. Chapter 12: An introduction to Time Series Analysis

Chapter 12: An introduction to Time Series Analysis. Chapter 12: An introduction to Time Series Analysis Chapter 12: An introduction to Time Series Analysis Introduction In this chapter, we will discuss forecasting with single-series (univariate) Box-Jenkins models. The common name of the models is Auto-Regressive

More information

Chapter 4: Models for Stationary Time Series

Chapter 4: Models for Stationary Time Series Chapter 4: Models for Stationary Time Series Now we will introduce some useful parametric models for time series that are stationary processes. We begin by defining the General Linear Process. Let {Y t

More information

Multivariate Time Series: VAR(p) Processes and Models

Multivariate Time Series: VAR(p) Processes and Models Multivariate Time Series: VAR(p) Processes and Models A VAR(p) model, for p > 0 is X t = φ 0 + Φ 1 X t 1 + + Φ p X t p + A t, where X t, φ 0, and X t i are k-vectors, Φ 1,..., Φ p are k k matrices, with

More information

ESSE Mid-Term Test 2017 Tuesday 17 October :30-09:45

ESSE Mid-Term Test 2017 Tuesday 17 October :30-09:45 ESSE 4020 3.0 - Mid-Term Test 207 Tuesday 7 October 207. 08:30-09:45 Symbols have their usual meanings. All questions are worth 0 marks, although some are more difficult than others. Answer as many questions

More information

Computational statistics

Computational statistics Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated

More information

Metric Spaces and Topology

Metric Spaces and Topology Chapter 2 Metric Spaces and Topology From an engineering perspective, the most important way to construct a topology on a set is to define the topology in terms of a metric on the set. This approach underlies

More information

Expressions for the covariance matrix of covariance data

Expressions for the covariance matrix of covariance data Expressions for the covariance matrix of covariance data Torsten Söderström Division of Systems and Control, Department of Information Technology, Uppsala University, P O Box 337, SE-7505 Uppsala, Sweden

More information

A time series is called strictly stationary if the joint distribution of every collection (Y t

A time series is called strictly stationary if the joint distribution of every collection (Y t 5 Time series A time series is a set of observations recorded over time. You can think for example at the GDP of a country over the years (or quarters) or the hourly measurements of temperature over a

More information

LECTURES 2-3 : Stochastic Processes, Autocorrelation function. Stationarity.

LECTURES 2-3 : Stochastic Processes, Autocorrelation function. Stationarity. LECTURES 2-3 : Stochastic Processes, Autocorrelation function. Stationarity. Important points of Lecture 1: A time series {X t } is a series of observations taken sequentially over time: x t is an observation

More information

Time Series Analysis -- An Introduction -- AMS 586

Time Series Analysis -- An Introduction -- AMS 586 Time Series Analysis -- An Introduction -- AMS 586 1 Objectives of time series analysis Data description Data interpretation Modeling Control Prediction & Forecasting 2 Time-Series Data Numerical data

More information

Ch. 15 Forecasting. 1.1 Forecasts Based on Conditional Expectations

Ch. 15 Forecasting. 1.1 Forecasts Based on Conditional Expectations Ch 15 Forecasting Having considered in Chapter 14 some of the properties of ARMA models, we now show how they may be used to forecast future values of an observed time series For the present we proceed

More information

Identifiability, Invertibility

Identifiability, Invertibility Identifiability, Invertibility Defn: If {ǫ t } is a white noise series and µ and b 0,..., b p are constants then X t = µ + b 0 ǫ t + b ǫ t + + b p ǫ t p is a moving average of order p; write MA(p). Q:

More information

Ross Bettinger, Analytical Consultant, Seattle, WA

Ross Bettinger, Analytical Consultant, Seattle, WA ABSTRACT DYNAMIC REGRESSION IN ARIMA MODELING Ross Bettinger, Analytical Consultant, Seattle, WA Box-Jenkins time series models that contain exogenous predictor variables are called dynamic regression

More information

7. Forecasting with ARIMA models

7. Forecasting with ARIMA models 7. Forecasting with ARIMA models 309 Outline: Introduction The prediction equation of an ARIMA model Interpreting the predictions Variance of the predictions Forecast updating Measuring predictability

More information

Lecture 2: Univariate Time Series

Lecture 2: Univariate Time Series Lecture 2: Univariate Time Series Analysis: Conditional and Unconditional Densities, Stationarity, ARMA Processes Prof. Massimo Guidolin 20192 Financial Econometrics Spring/Winter 2017 Overview Motivation:

More information

Module 4. Stationary Time Series Models Part 1 MA Models and Their Properties

Module 4. Stationary Time Series Models Part 1 MA Models and Their Properties Module 4 Stationary Time Series Models Part 1 MA Models and Their Properties Class notes for Statistics 451: Applied Time Series Iowa State University Copyright 2015 W. Q. Meeker. February 14, 2016 20h

More information

Lecture 3-4: Probability models for time series

Lecture 3-4: Probability models for time series Lecture 3-4, page 1 Lecture 3-4: Probability models for time series Outline of lesson 3-4 (chapter 3) The most heavy (and theoretical) part (and lesson) AR(p) processes MA(q) processes ARMA(p,q) processes

More information

Forecasting with ARMA

Forecasting with ARMA Forecasting with ARMA Eduardo Rossi University of Pavia October 2013 Rossi Forecasting Financial Econometrics - 2013 1 / 32 Mean Squared Error Linear Projection Forecast of Y t+1 based on a set of variables

More information

Chapter 3. ARIMA Models. 3.1 Autoregressive Moving Average Models

Chapter 3. ARIMA Models. 3.1 Autoregressive Moving Average Models Chapter 3 ARIMA Models Classical regression is often insu cient for explaining all of the interesting dynamics of a time series. For example, the ACF of the residuals of the simple linear regression fit

More information

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY Time Series Analysis James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY PREFACE xiii 1 Difference Equations 1.1. First-Order Difference Equations 1 1.2. pth-order Difference Equations 7

More information

Trend-Cycle Decompositions

Trend-Cycle Decompositions Trend-Cycle Decompositions Eric Zivot April 22, 2005 1 Introduction A convenient way of representing an economic time series y t is through the so-called trend-cycle decomposition y t = TD t + Z t (1)

More information

INTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING

INTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING INTRODUCTION TO MARKOV CHAINS AND MARKOV CHAIN MIXING ERIC SHANG Abstract. This paper provides an introduction to Markov chains and their basic classifications and interesting properties. After establishing

More information

Multivariate Time Series

Multivariate Time Series Multivariate Time Series Notation: I do not use boldface (or anything else) to distinguish vectors from scalars. Tsay (and many other writers) do. I denote a multivariate stochastic process in the form

More information

STOCHASTIC PROCESSES Basic notions

STOCHASTIC PROCESSES Basic notions J. Virtamo 38.3143 Queueing Theory / Stochastic processes 1 STOCHASTIC PROCESSES Basic notions Often the systems we consider evolve in time and we are interested in their dynamic behaviour, usually involving

More information

Ch. 14 Stationary ARMA Process

Ch. 14 Stationary ARMA Process Ch. 14 Stationary ARMA Process A general linear stochastic model is described that suppose a time series to be generated by a linear aggregation of random shock. For practical representation it is desirable

More information

Strictly Stationary Solutions of Autoregressive Moving Average Equations

Strictly Stationary Solutions of Autoregressive Moving Average Equations Strictly Stationary Solutions of Autoregressive Moving Average Equations Peter J. Brockwell Alexander Lindner Abstract Necessary and sufficient conditions for the existence of a strictly stationary solution

More information

A quick introduction to Markov chains and Markov chain Monte Carlo (revised version)

A quick introduction to Markov chains and Markov chain Monte Carlo (revised version) A quick introduction to Markov chains and Markov chain Monte Carlo (revised version) Rasmus Waagepetersen Institute of Mathematical Sciences Aalborg University 1 Introduction These notes are intended to

More information

Ch 6. Model Specification. Time Series Analysis

Ch 6. Model Specification. Time Series Analysis We start to build ARIMA(p,d,q) models. The subjects include: 1 how to determine p, d, q for a given series (Chapter 6); 2 how to estimate the parameters (φ s and θ s) of a specific ARIMA(p,d,q) model (Chapter

More information

EC402: Serial Correlation. Danny Quah Economics Department, LSE Lent 2015

EC402: Serial Correlation. Danny Quah Economics Department, LSE Lent 2015 EC402: Serial Correlation Danny Quah Economics Department, LSE Lent 2015 OUTLINE 1. Stationarity 1.1 Covariance stationarity 1.2 Explicit Models. Special cases: ARMA processes 2. Some complex numbers.

More information

STA 6857 Autocorrelation and Cross-Correlation & Stationary Time Series ( 1.4, 1.5)

STA 6857 Autocorrelation and Cross-Correlation & Stationary Time Series ( 1.4, 1.5) STA 6857 Autocorrelation and Cross-Correlation & Stationary Time Series ( 1.4, 1.5) Outline 1 Announcements 2 Autocorrelation and Cross-Correlation 3 Stationary Time Series 4 Homework 1c Arthur Berg STA

More information

Econ 623 Econometrics II Topic 2: Stationary Time Series

Econ 623 Econometrics II Topic 2: Stationary Time Series 1 Introduction Econ 623 Econometrics II Topic 2: Stationary Time Series In the regression model we can model the error term as an autoregression AR(1) process. That is, we can use the past value of the

More information

Some Time-Series Models

Some Time-Series Models Some Time-Series Models Outline 1. Stochastic processes and their properties 2. Stationary processes 3. Some properties of the autocorrelation function 4. Some useful models Purely random processes, random

More information

Statistics of stochastic processes

Statistics of stochastic processes Introduction Statistics of stochastic processes Generally statistics is performed on observations y 1,..., y n assumed to be realizations of independent random variables Y 1,..., Y n. 14 settembre 2014

More information

LECTURE 10 LINEAR PROCESSES II: SPECTRAL DENSITY, LAG OPERATOR, ARMA. In this lecture, we continue to discuss covariance stationary processes.

LECTURE 10 LINEAR PROCESSES II: SPECTRAL DENSITY, LAG OPERATOR, ARMA. In this lecture, we continue to discuss covariance stationary processes. MAY, 0 LECTURE 0 LINEAR PROCESSES II: SPECTRAL DENSITY, LAG OPERATOR, ARMA In this lecture, we continue to discuss covariance stationary processes. Spectral density Gourieroux and Monfort 990), Ch. 5;

More information

Minicourse on: Markov Chain Monte Carlo: Simulation Techniques in Statistics

Minicourse on: Markov Chain Monte Carlo: Simulation Techniques in Statistics Minicourse on: Markov Chain Monte Carlo: Simulation Techniques in Statistics Eric Slud, Statistics Program Lecture 1: Metropolis-Hastings Algorithm, plus background in Simulation and Markov Chains. Lecture

More information

Calculation of ACVF for ARMA Process: I consider causal ARMA(p, q) defined by

Calculation of ACVF for ARMA Process: I consider causal ARMA(p, q) defined by Calculation of ACVF for ARMA Process: I consider causal ARMA(p, q) defined by φ(b)x t = θ(b)z t, {Z t } WN(0, σ 2 ) want to determine ACVF {γ(h)} for this process, which can be done using four complementary

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

Ch 5. Models for Nonstationary Time Series. Time Series Analysis

Ch 5. Models for Nonstationary Time Series. Time Series Analysis We have studied some deterministic and some stationary trend models. However, many time series data cannot be modeled in either way. Ex. The data set oil.price displays an increasing variation from the

More information

Time Series: Theory and Methods

Time Series: Theory and Methods Peter J. Brockwell Richard A. Davis Time Series: Theory and Methods Second Edition With 124 Illustrations Springer Contents Preface to the Second Edition Preface to the First Edition vn ix CHAPTER 1 Stationary

More information

Chapter 9: Forecasting

Chapter 9: Forecasting Chapter 9: Forecasting One of the critical goals of time series analysis is to forecast (predict) the values of the time series at times in the future. When forecasting, we ideally should evaluate the

More information

Università di Pavia. Forecasting. Eduardo Rossi

Università di Pavia. Forecasting. Eduardo Rossi Università di Pavia Forecasting Eduardo Rossi Mean Squared Error Forecast of Y t+1 based on a set of variables observed at date t, X t : Yt+1 t. The loss function MSE(Y t+1 t ) = E[Y t+1 Y t+1 t ]2 The

More information