Model Identification for Time Series with Dependent Innovations

Size: px
Start display at page:

Download "Model Identification for Time Series with Dependent Innovations"

Transcription

1 Model Identification for Time Series with Dependent Innovations Shuhao Chen 1, Wanli Min 2 and Rong Chen 1,3 1 Rutgers University, 2 IBM Singapore and 3 Peking University Abstract: This paper investigates the impact of dependent but uncorrelated innovations (errors) on the traditional autoregressive moving average model (ARMA) order determination schemes such as autocorrelation function (ACF), partial autocorrelation function (PACF), extended autocorrelation function (EACF) and unit-root test. The ARMA models with iid innovations have been studied extensively and are well-posed, but their properties with dependent but uncorrelated innovations are relatively less studied. In the presence of such innovations, we show that the ACF, PACF and EACF are significantly impacted while the unitroot test is not affected. We also propose a new order determination scheme to address those impacts for analyzing time series with uncorrelated innovations. Key words and phrases: time series, order determination, uncorrelated but dependent errors, ACF, PACF, EACF, unit-root test, GARCH. 1 Introduction Time series analysis is widely used in many fields including econometrics, finance, engineering and metrology. The most commonly used time series models are the autoregressive moving average (ARMA) model. More specifically, ARM A(p, q) model for a univariate time series X t takes the form of Φ(B)X t = Ψ(B)ε t, (1) with Φ(B) = φ(b)u(b) where Φ(B) = 1 Φ 1 B Φ p B p, U(B) = 1 U 1 B U d B d, φ(b) = 1 φ 1 B φ p d B p d and Ψ(B) = 1+ψ 1 B+ +ψ q B q are polynomials in the back-shift operator B, defined as BX t = X t 1 and Bε t = ε t 1. We shall require that all zeros of U(B) are on the unit circle and those of φ(b) and Ψ(B) are outside the unit circle. ε t is the innovation term. Note that when U(B) = (1 B) d, model (1) is autoregressive moving integrated moving average ARIMA(p d, d, q) model. If we further assume Ψ(B) = 1, model (1) is ARI(p d, d) model. 1

2 The first step and also one of the key steps in building a time series model is order determination. In the literature, order determination schemes for time series models with identically independent distributed (iid) innovations{ε t } are well studied. Box and Jenkins (1976) introduced the autocorrelation function (ACF) and partial autocorrelation function (PACF). The Akaike Information Criterion (AIC) by Akaike (1974) and Bayesian Information Criteria (BIC) by Schwarz (1978) are two goodness of fit measures of an estimated model to facilitate the model selection. Tsay and Tiao (1984) proposed the extended autocorrelation function (EACF) for order determination of ARM A(p, q) model. On the other hand, Dickey and Fuller (1979) studied the unit-root behavior and gave the asymptotic distribution of a unit root test statistic. Standard order determination procedure combines those two techniques: taking the unit root test to decide whether to make difference(s)(i.e. set Y t = X t X t 1 ) and then using ACF/PACF/EACF procedure on differenced series Y t to get AR and MA orders p and q respectively. Other models and order determination schemes include R and S array approach by Gray, Kelley and McIntire (1978); Corner method by Beguin, Gourieroux and Monfort (1980) and Smallest Canonical Correlation (SCAN) by Tsay and Tiao (1985). Most of the existing order determination methods assume that the innovation sequence {ε t }, are iid and/or have constant conditional variance, i.e., E[ε 2 t F t 1 ] = σ 2, excluding many interesting processes such as the generalized autoregressive conditional heteroscedastic (GARCH) models by Bollerslev (1986) and the stochastic volatility (SV) model by Melino and Turnbull (1990), both often used in financial time series modeling. In these models, the innovation series are not auto-correlated but auto-dependent. Thus it becomes interesting and important to revisit the classical order determination schemes at the presence of auto-dependent but uncorrelated innovations. Specifically, in this paper we study the order determination schemes for MA, AR, ARMA and ARI models at the presence of uncorrelated but dependent innovations. Autocorrelation function (ACF) is a simple order determination procedure for MA process and for AR process partial autocorrelation function (PACF) becomes effective. For ARMA process, extended autocorrelation function (EACF) has shown to be very useful in identifying the AR and MA orders and it also works for differenced sequence of ARIMA process. We investigate how those schemes are impacted by the new type of innovations considered here. There have been several studies in this area. Min (2004) investigated the 2

3 effect of dependent innovations on pure AR and MA time series sequences; Yang and Zhang (2008) discussed the unit-root test with GARCH(1,1) innovations, one of the most used dependent but uncorrelated innovations. Our study is under a more unified framework which allows for a wide class of uncorrelated but dependent innovations. The rest of the paper is organized as follows. In section 2, we present a detailed study on how dependent but uncorrelated innovations affect the properties of the key statistics in classical time series model identification procedures. Based on these findings, we propose suitable methods and establish their theoretical validations. Simulation studies and applications to real examples are illustrated in section 3. Proofs are given in the Appendix. For simplicity, we employ the following notations and terms throughout the dist P content: means convergence in distribution; means convergence in probability; a.s. means almost sure convergence; = means denotation; term errors and innovations are used interchangeably. For a set of random variables {X n } and a corresponding set of function {f n }, the notation X n = O p (f n ) means that the set of values X n /f n is bounded in the limit in probability. 2 Method In this paper we consider the following structure of dependent but uncorrelated innovations. Let {e t } be iid random variables and F be a measurable function such that innovation ε t = F (e t, e t 1, ) is a well-defined random variable. We assume the following condition. Condition C.1: E(ε t F t 1 ) = 0 and E(ε 2 t ) = σ 2, t Z, where F t = σ(et, e t 1, ) is the σ-field generated by sequence {e t } representing the information available up to time t. The error series {ε t } is uncorrelated since E(ε t ε t 1 ) = E(E(ε t ε t 1 F t 1 )) = E(ε t 1 E(ε t F t 1 )) = 0. The second part of condition C.1 is weaker than the traditional condition E(ε 2 t F t 1 ) = σ 2, which can be seen from the iterative expectation E(ε 2 t ) = E(E(ε 2 t F t 1 )). The stronger condition E(ε 2 t F t 1 ) = σ 2 also implies zero autocorrelations of sequence {ε 2 t }, i.e., Cov(ε 2 t, ε 2 t 1 ) = 0. However, condition C.1 allows nonzero autocorrelations of {ε 2 t }. Another important characterization of the uncorrelated and dependent innovation {ε t } is related to projections defined in Wu and Min (2005): P t η = 3

4 E(η F t ) E(η F t 1 ). For instance, P 1 ε i ε j = E(εi ε j F 1 ) E(ε i ε j F 0 ). Denote and p as L 2 and L p norm, respectively. We assume the following condition. Condition C.2: E(ε 4 n) < and P 1 ε t 4 <, t=1 i,j=0 P 1ε i ε j <. Remark 1 The intuition of condition C.2 is that the projection of the future ε i ε j to the space µ 1 µ 0 = {Z : i,j=0 P 1Z <, Z is F 1 measurable and E(Z F 0 ) = 0} has a small magnitude; namely, the future depends weakly on the current states or the current states depends weakly on the previous states. This setup includes many interesting stochastic processes, such as ARCH(p) model, GARCH(p, q) model, and Stochastic Volatility Model. For example, Min (2004) showed that GARCH(p, q) model satisfies condition C Moving Average Model One of the fundamental building blocks of time series models is the moving average (MA) model. Specifically, a moving average process X t with order q, namely MA(q) is defined as follows X t = ε t + ψ 1 ε t ψ q ε t q. (2) It is a special case of ARMA model (1) with Φ(B) = 1. The autocovariance function and autocorrelation function (ACF) are defined as γ(h) = Cov(x t, x t+h ) and ρ(h) = Cor(x t, x t+h ), (3) respectively. The ACF has the unique feature that for an M A(q) process, its ACF cuts off at lag q. Since {ε t } are uncorrelated, we have γ(h) = σ 2 q i=0 ψ iψ i+h, where ψ j = 0 for j > q. Hence γ(q + i) = ρ(q + i) = 0, i > 0. ACF can be used effectively to determine the order q for an MA(q) model. Let x 1,, x n be a sample of MA(q) process in (2). We denote the sample autocovariance function and sample autocorrelation function respectively as ˆγ(h) = 1 n h x t x t+h and ˆρ(h) = ˆγ(h)/ˆγ(0). (4) n t=1 It is known (e.g. Brockwell and Davis, 1986) that n(ˆρ(h) ρ(h)) dist 4

5 N(0, W ) where W is given by the celebrated Bartlett formula W = k= {ρ 2 (k) + ρ(h k)ρ(h + k) + 2ρ 2 (h)ρ 2 (k) 4ρ(h)ρ(k)ρ(k h)}. (5) If we further assume ε t iid (0, σ 2 ), the Bartlett formula (5) implies that for M A(q) model, W = (1 + 2ρ 2 (1) + + 2ρ 2 (q)), if h > q. However, when {ε t } are dependent, the validity of using n 1 W to estimate the variance of sample ACF needs to be reexamined. In the following, we establish asymptotic property of ˆρ(h) under conditions C.1 and C.2 of {ε t }. Specifically, we show that the cut-off property (Lemma 1) and the asymptotic joint normality of ˆγ(h) still holds, but the asymptotic variance is different (Theorem 1). Lemma 1 Let {X t } be a MA(q) process with innovation {ε t } following conditions C.1. For the autocorrelation function ρ(h) defined in (3), we have ρ(q + 1) = ρ(q + 2) = = 0. The proof is trivial since the cut-off properties is only related to correlation rather than dependence. Theorem 1 Let {X t } be a MA(q) process with innovation {ε t } following conditions C.1 and C.2; autocorrelation function ρ(h) defined as (3) and sample autocorrelation ˆρ(h) defined as (4), then we have h 1, ( ) dist n(ˆρ(1) ρ(1),, ˆρ(h) ρ(h)) N 0, Σ, h 1 h h where Σ assumes a complex form to be explicitly expressed in the proof in Appendix. In particular, for h > q, Var(ˆρ(h)) 1 ( ) σ(0, h) + 2σ(1, h) + + 2σ(q, h), (6) nγ 2 (0) where σ(d, h) = E(x 0 x h x d x d+h ). If it is further assumed that ACF of {ε 2 t } to be 5

6 nonnegative and E(ε 2 t ε t i ε t j ) = 0, i j, then for any h > q, nˆρ(h) dist N(0, δ 2 h ), q δ2 h (1 + 2 ρ 2 (k)). (7) k=1 The inequality holds strictly if the ACFs are positive. The proof of the theorem is in the Appendix. Remark 2 The theorem applies to the finite 4 th moment GARCH(p, q) process and stochastic volatility models (see definition in (15)) as they satisfy the conditions of the theorem and additional positive ACF conditions. Based on the theorem, one can estimate the standard error of ˆρ(h) by (6), with σ(d, h) replaced by its estimate. However, the order of the MA process q is unknown. Hence, we use the following estimator, Var( ˆρ(h)) = 1 ) (ˆσ(0, h) + 2ˆσ(1, h) + + 2ˆσ(h 1, h). (8) nˆγ 2 (0) This is due to the fact that E(x t x t+h x s x s+h ) = 0 for s t > q for a MA(q) process, hence we have σ(q + 1, h) = σ(q + 2, h) = = σ(h 1, h) = 0 for h > q. Thus including ˆσ(q + 1, h) = ˆσ(q + 2, h) = = ˆσ(h 1, h) would not influence the estimator of the asymptotic variance. In section 3.1.1, we perform a comparison study between the variance calculation above and the standard variance formula 1 n (1+2ˆρ2 (1)+ +2ˆρ 2 (h 1)) for i.i.d innovations. Because of the differences in the variance, order identifications via ACF needs to be adjusted. A demonstration is represented in section It is noted that although our estimator is built for dependent but uncorrelated innovations, it applies to iid innovations automatically. 2.2 Autoregressive Model An autoregressive process X t with order p, namely AR(p), follows X t = φ 1 X t 1 + φ 2 X t φ p X t p + ε t. It is a common practice to test the significance of the sample partial autocorrelation function (PACF) for identifying the order of AR(p) series. The PACF φ h,h 6

7 is defined as φ 0,0 = 1 and φ h,h = the last component of φ h = Γ 1 h γ h, h 1, (9) where Γ h = [γ(i j)] h i,j=1 and γ h = (γ(1), γ(2),, γ(h)). It is estimated by replacing γ(i) with its estimate ˆγ(i) in the above formula. It can be shown (e.g. Brockwell and Davis, 1986) that the partial autocorrelation φ h,h at lag h may be regarded as the correlation between X t and X t h, adjusted for the intervening observations X t 1,, X t h+1. Since an AR(p) model imposes a linear relationship between X t and X t 1,, X t p, it is easily seen that when h > p, X t and X t h are conditionally uncorrelated, given X t 1,, X t h+1, hence the cut-off property of PACF of AR processes. If one further assumes Gaussian white noise, i.e., ε t iid N(0, σ 2 ), it is well known that h > p, denote s = h p 1, n( ˆφ p+1,p+1,, ˆφ p+s,p+s ) has an asymptotic joint normal distribution with unit variance. Namely, we have n( ˆφp+1,p+1,, ˆφ ( ) p+s,p+s ) dist N 0, I, (10) s 1 s s where I is the s dimensional identity matrix. However, similar to the MA(q) case, when {ε t } are dependent, the variance matrix of the asymptotic distribution of PACF is different from the iid case. Lemma 2 Assume {X t } to be an AR(p) process with innovation {ε t } following condition C.1. For the PACF φ h,h defined in (9), we have φ p+1,p+1 = φ p+2,p+2 = = 0. In analogy to (10), the next theorem establishes the asymptotic properties of sample PACF ˆφ h,h, h > p of an AR(p) process with dependent innovations. Theorem 2 Let ˆφ h,h be the lag h sample PACF of a stationary and invertible AR(p) time series {X t }, where the innovations {ε t } satisfies conditions C.1 and C.2. Then h > p (s = h p 1), n( ˆφp+1,p+1,, ˆφ ( ) p+s,p+s ) dist N 0, Ξ. s 1 s s where Ξ has a complex expression to be given in the proof of the theorem in Ap- 7

8 pendix. If we further assume ACF of {ε 2 t } to be nonnegative and E(ε 2 t ε t i ε t j ) = 0, i j, then. Ξ (l,l) 1, l (1,, s). (11) The strict inequality holds if all ACFs are positive. Remark 3 The dependent but uncorrelated innovations such as GARCH innovations render a large variance of sample PACF compared with the case of iid innovations under which the asymptotic variance is 1 n. So if we still use the traditional (1 α)% confidence interval to test for lag-h PACF with h > p, the type I error would be larger than α%. This over rejection leads to a specification of AR(h) model with h greater than the true order p. We can obtain the variance estimator of the sample lag-h PACF Var( ˆφ h,h ) in the following manner. First we obtain the estimate ˆφ 1,, ˆφ h and the corresponding residuals ˆε t = x t ˆφ 1 x t 1 ˆφ h x t h. Secondly, from the sample autocovariance matrix ˆΓ h and the sample version of Ω h = Cov(ε1 x 0, ε 1 x 1,, ε 1 x 1 h ), we obtain an estimate of Γ 1 h Ω hγ 1 h whose last diagonal entry over n is the variance estimator. See proof of Theorem 2 in Appendix for more details. In section 3.1.2, we perform a comparison study between variance calculation above and the standard variance 1 n under i.i.d case. It demonstrates that the order identification scheme for AR(p) process using PACF needs to be adjusted to the updated variance. The illustration of order identifications via PACF corresponding to different innovations is represented in section Autoregressive and Moving Average Model ARMA model in (1) is a hybrid of autoregressive model and moving average model. Tsay and Tiao (1984) proposed the extended autocorrelation function (EACF) technique to identify the orders of a stationary or non-stationary ARMA process based on iterated least square estimates of the autoregressive parameters. It is based on the fact that, if X t follows an ARMA(p, q) model, then the filtered process Y t = X t φ 1 X t 1 φ p X t p follows an MA(q) process Y t = ε t + ψ 1 ε t ψ q ε t q. 8

9 Under the dependent but uncorrelated innovations, in analogy to the procedure described in Tsay and Tiao (1984), the order determination scheme is proposed as follows. 1. For each candidate AR order s, we first obtain a consistent estimate of the AR coefficients, ˆφ 1, ˆφ 2,, ˆφ s. 2. Obtain Y t = X t ˆφ 1 X t 1 ˆφ 2 X t 2 ˆφ s X t s. If s = p, then Y t should approximately follow an M A(q) sequence. On the other hand, if we overfit the AR term by m (i.e. s = p + m), then Y t should be an MA(q + m) process. 3. Use the order determination procedures for moving average sequences described above on {Y t } to identify the MA order and build an EACF table as proposed in Tsay and Tiao (1984) and finalize the AR and MA order selections. Since the major difference of this scheme with that of Tsay and Tiao (1984) is in the ways of dealing with the significance of ACF elaborated in Section 2.1, we skip the discussion here. The impact on EACF process inference of different innovation terms is illustrated in Section Autoregressive Integrated Model Autoregressive integrated (ARI) model can be regarded as an extension of the AR model. The Augmented Dickey Fuller (ADF) test (Dickey & Fuller, 1979, Said and Dickey, 1984) is important in detecting whether the process is unit-root nonstationary. In this section we study ADF under uncorrelated but dependent innovations. Consider ARI model, X t = Φ 1 X t 1 + Φ 2 X t Φ p X t p + ε t. (12) Following the formulation of ADF, we present (12) as p 1 X t = ρx t 1 + a k X t k + ε t, (13) k=1 where X t k = Xt k X t k 1, ρ = p i=1 Φ i and a k = p j=k+1 Φ j. The 9

10 unit-root problem for this model focuses on testing H 0 : ρ = 1 against H 1 : ρ < 1. The ADF test statistic is T n = n( ρn 1), where ρ n is the least square estimator of ρ when X t is regressed on X t 1, X t 1,, X t p+1 using (13), ρ n = n t=p+1 X t 1(X t p 1 k=1 âk X t k ). n t=p+1 X2 t 1 Dickey and Fuller (1979) proved that under the null hypothesis T n = n( ρ n 1) dist 1 p 1 2 (1 a k )(Γ 1 (T 2 1)), where (Γ, T ) = ( i=1 γ2 i Z2 i, i=1 2γi Z i ), γ 2 i = 4[(2i 1)π] 2 and {Z i } i=1 is a sequence of iid N(0, 1) random variables. Chan and Wei (1988) showed that the limiting distribution could also be further represented in terms of function of Brownian Motion W (t), k=1 T n = n( ρ n 1) dist 1 p 1 2 (1 a k ) (W 2 (1) 1) 1 k=1 0 W 2 (t)dt. When the innovation {ε t } is not a sequence of iid random variables but dependent and uncorrelated random variables satisfying conditions C.1 and C.2, we show in the following that the same limiting distribution of the test statistics T n still applies. Before we state the results, we need the following definition from Wu and Min (2005) and lemmas. Definition 1 The process Y n = g(f n ), where g is a measurable function, is said to be L p (p 0) weakly dependent with order r(r 0) if E( Y n p ) < and n r P 1 Y n p <. (14) n=1 If (14) holds with r = 0, then Y n is said to be L p weakly dependent. We impose an additional condition on the innovation {ε t }. 10

11 Condtion C.3: The {ε t } process is L α is weakly dependent with order 1, for some α > 2. This innovation assumption includes a large class of nonlinear processes, and substantially relaxes the iid or martingale difference assumption. In particular, this condition is satisfied if the innovations are GARCH process, random coefficient AR, bilinear AR, etc. For more details, see Wu and Min (2005). The next theorem indicates that the limiting distribution of the test statistic T n does not change if we use dependent but uncorrelated errors. Theorem 3 If ρ = 1, for time series {X t } following (12) and errors {ε t } under the conditions C.1, C.2. and C.3, we have W n dist W, and therefore (1 a k ) T n dist p 1 k=1 1 0 W (t)dw (t) 1 0 W. 2 (t)dt Remark 4 According to Theorem of Fuller (1995), â k is a consistent estimator of a k and we can use â k in estimating the limiting distribution, where (ˆρ, â 1,, â k ) is the regression coefficients obtained by regressing X t on X t 1, X t 1,, X t p+1. Remark 5 Theorem 3 can not be easily extended to ARIMA models since equation (12) is no longer a simple regression model when MA part is involved. Further study would be performed on this topic. 3 Empirical Study In this section, we use both simulation and real data to demonstrate the results and methods we developed in the previous section. Simulation results for AR/MA/ARMA/ARI respectively are provided first to validate the theorems. Only part of the results are presented here. Results with different parameters of the GARCH model and results with other innovation models such as stochastic volatility model would be given upon request. For real data sets, we go through the entire order determination and estimation process. We will denote the EACF proposed by Tsay and Tiao (1984) as original EACF and the adjusted EACF for 11

12 time series associated with uncorrelated and dependent innovations as modified EACF. 3.1 Simulation Here we use GARCH(1, 1) innovations as the dependent but uncorrelated errors since GARCH is among the most common and useful models for this type of errors. Namely, we have the following model, { X t = φ 1 X t φ p X t p + ε t + + ψ q ε t q, ε t = g t ɛ t, g t = α 0 + αε 2 t 1 + βg t 1, where {ɛ t } are iid normal random sequences Moving Average Model We consider two formulae for estimating Var( ˆρ(h)): 1. V (h) = 1 n (1 + 2ˆρ2 (1) + + 2ˆρ 2 (h 1)); 2. V (h) = 1 (ˆσ(0, h) + 2ˆσ(1, h) + + 2ˆσ(h 1, h)). nˆγ 2 (0) Correspondingly, we denote the statistics ˆρ 2 (h)/v (h) and ˆρ 2 (h)/v (h) as T (h) and T (h) respectively. First, we use the following M A(1) model for simulation. { X t = ε t + ψ 1 ε t 1, ε t = g t ɛ t, g t = α 0 + αε 2 t 1 + βg t 1, with ψ 1 = 0.4 and GARCH effect setting: α = 0.2, β = 0.7. We simulate 2000 replications of time series with length 1000 each. In the Table 1, we report the empirical percentiles of T (h) and T (h) along with that of χ 2 1, for h = 2 and 3. Since the series are simulated from MA(1), the true value of ρ should be zero and asymptotically the statistics should follow χ 2 1 if the variance is estimated correctly. The last column shows the percentage of rejections if χ 2.95 is used as the critical value for testing non-zero ACF coefficients. It can be seem from Table 1 that T (h) closely follows the chi-square distribution while T (h) is much larger. Since the numerate ˆρ(h) is the same for both T (h) and T (h), it shows that the variance estimate V (h) is more accurate. The level of testing using T (h) is also more accurate. 12

13 Table 1: Empirical percentile of T (h) and T (h) for a simulated MA(1) series X t = ε t 0.4ε t 1 where innovation {ε t } is GARCH(1,1) with α = 0.2, β = 0.7. Mean S.D. 50% 75% 90% 95% 99% p T (2) % T (3) % χ T (2) % T (3) % In order to check the property of T (h) and T (h) when h q, we use the following M A(3) model for a second simulation. { X t = ε t + ψ 1 ε t 1 + ψ 2 ε t 2 + ψ 3 ε t 3, ε t = g t ɛ t, g t = α 0 + αε 2 t 1 + βg t 1, with ψ 1 = 0.8, ψ 2 = 0.8, ψ 3 = 0.8 and two different GARCH effect settings: i) α = 0.1, β = 0.8; ii) α = 0.5, β = 0.2. Again, we simulate 2000 replications of the time series with length 1000 each. The empirical percentiles of T (h) and T (h) with those of χ 2 1 are compared in the Table 2. Table 2: Empirical percentile of T (h) and T (h) for a simulated MA(3) series X t = ε t + 0.8ε t 1 0.8ε t ε t 3 where innovation ε t is GARCH(1,1) with α = 0.5, β = 0.2. Mean S.D. 50% 75% 90% 95% 99% p T (2) % T (3) % T (4) % T (5) % χ T (2) % T (3) % T (4) % T (5) % We make the following observations. 1. For h q, both T (h) and T (h) are significantly different from the χ

14 since ˆρ 2 (h)/ Var(ˆρ(h)) is asymptotically non-central χ 2 distributed as the asymptotic mean of ˆρ(h) is ρ(h) 0 when h q. 2. T (3) and T (3) are relatively greater than the others simply because ρ(3) is of a larger value. Theoretically, ρ(1) = 0.164, ρ(2) = 0.055, ρ(3) = 0.274, ρ(4) = ρ(5) = For h > q, the performance of T (h) and T (h) are similar to what we observed in MA(1) simulation. T (h) follows the χ 2 1 distribution closely while T (h) is much larger. 4. It is clear that while the power of testing using T (h) is comparable with that of T (h), its size is more accurate. Since order determination is based on the change from non-zero ACF coefficients to zero ACF coefficients, T (h) is certainly more reliable. Having confirmed that the variance of ˆρ(h) may change when faced with dependent innovations, we need to use the new order identification scheme discussed in previous section. Since MA model is a specific case of ARMA model, we incorporate the interpretation of the results of the new scheme compared to the classical one into our discussion of the ARMA model and only list the result table here (Table 3). All the interpretation in Section can be applied here though Autoregressive Model For comparing the variance estimator under i.i.d assumption and our proposed estimator, we consider two formulae for estimating the variance Var( ˆφ h,h ): 1. V (h, h) = 1 n ; 2. V (h, h) = 1 n [ˆΓ 1 h ˆΩ hˆγ 1 h ] (h,h). Correspondingly, we denote the statistics ˆφ 2 h,h /V (h, h) and ˆφ 2 h,h /V (h, h) as T (h) and T (h) respectively. We use AR(1) model for the first simulation experiment in this setting. { X t = φ 1 X t 1 + ε t, ε t = g t ɛ t, g t = α 0 + αε 2 t 1 + βg t 1, with φ 1 = 0.9 and GARCH effect setting: α = 0.5, β =

15 Table 3: EACF method comparison for a simulated MA(3) series with different type of innovations: X t = ε t + 0.8ε t 1 0.8ε t ε t 3 where innovation ε t s types are: (1) iid; (2) GARCH(1,1) with α = 0.1, β = 0.8; (3) GARCH(1,1) with α = 0.5, β = 0.2. The time series length is 1000 while the replication is 1000 as well. The numbers (in %) represent the percentage which the order (in row) identified by the method (in column) for the simulated series. Each column would sum up to 100%. Innovation (1) Innovation (2) Innovation (3) Models original modified original modified original modified MA(2) 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% MA(3) 85.10% 89.80% 76.50% 87.70% 69.40% 85.00% MA(4) 6.60% 4.10% 9.30% 5.90% 13.90% 5.10% MA(5) 2.60% 1.90% 6.10% 2.20% 6.10% 1.90% ARMA(1,3) 0.20% 0.00% 0.30% 0.10% 0.40% 0.20% ARMA(1,4) 2.10% 1.60% 3.00% 0.50% 3.80% 0.00% ARMA(1,5) 0.00% 0.00% 0.00% 0.00% 0.00% 1.00% ARMA(2,3) 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% ARMA(2,4) 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% ARMA(2,5) 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% ARMA(3,3) 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% ARMA(3,4) 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% Others 3.40% 2.60% 4.80% 3.60% 6.40% 6.80% 15

16 We simulate 2000 replications of time series with length 1000 each. The empirical percentiles of T (h) and T (h) with those of χ 2 1 are listed in the Table 4. It shows very similar results as that of Table 1 and leads to similar conclusions. Table 4: Empirical percentile of T (h) and T (h) for a simulated AR(1) series X t = 0.9X t 1 + ε t where innovation {ε t } is GARCH(1,1) with α = 0.5, β = 0.2. Mean S.D. 50% 75% 90% 95% 99% p T (2) % T (3) % χ T (2) % T (3) % For h p, we use the following AR(3) model for a second simulation. { X t = φ 1 X t 1 + φ 2 X t 2 + φ 3 X t 3 + ε t, ε t = g t ɛ t, g t = α 0 + αε 2 t 1 + βg t 1, with φ 1 = 0.8, φ 2 = 0.8, φ 3 = 0.8 and two different GARCH effect settings: i) α = 0.1, β = 0.8; ii) α = 0.5, β = 0.2. Again, we simulate 2000 replications of time series with length 1000 each. The empirical percentiles of T (h) and T (h) with those of χ 2 1 are compared in the Table 5. Again, we see very similar results as those in Table 2. Table 5: Empirical percentile of T (h) and T (h) for a simulated AR(3) series X t = 0.8X t 1 0.8X t X t 3 + ε t where innovation ε t is GARCH(1,1) with α = 0.5, β = 0.2. Mean S.D. 50% 75% 90% 95% 99% p T (2) % T (3) % T (4) % T (5) % χ T (2) % T (3) % T (4) % T (5) % 16

17 Table 6: EACF method comparison for a simulated AR(3) series with different type of innovations: X t = 0.8X t 1 0.8X t X t 3 + ε t where innovation ε t s types are: (1) iid; (2) GARCH(1,1) with α = 0.1, β = 0.8; (3) GARCH(1,1) with α = 0.5, β = 0.2. The time series length is 1000 while the replication is 1000 as well. The numbers (in %) represent the percentage which the order (in row) identified by the method (in column) for the simulated series. Each column would sum up to 100%. Innovation (1) Innovation (2) Innovation (3) Models original modified original modified original modified AR(2) 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% ARMA(2,1) 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% ARMA(2,2) 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% AR(3) 57.90% 64.80% 48.40% 69.20% 33.60% 70.60% ARMA(3,1) 8.50% 10.40% 10.00% 8.60% 16.50% 7.40% ARMA(3,2) 5.30% 5.00% 7.70% 4.10% 10.10% 3.90% AR(4) 0.70% 0.60% 0.90% 0.50% 0.60% 1.00% ARMA(4,1) 18.20% 13.70% 19.30% 13.10% 18.20% 12.60% ARMA(4,2) 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% AR(5) 1.20% 0.30% 1.30% 0.30% 1.60% 1.10% ARMA(5,1) 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% ARMA(5,2) 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% Others 8.20% 5.20% 12.40% 4.20% 19.40% 3.40% Having confirmed that the variance of ˆφ h,h may change when faced with dependent innovations, we need to use the new order identification scheme discussed in previous section. Since AR model is also a specific case of ARMA model, we incorporate the interpretation of the results of the new scheme compared to the classical one into our discussion of the ARMA model and only report the results in Table 6. All the interpretation in section can be applied here though Autoregressive Moving Average Model Here we compare the different inferences of EACF, the original EACF and the modified EACF, on time series with either iid innovations or uncorrelated but dependent innovations. The experiment is designed as follows. 1. Simulate ARM A(1, 1) time series with three different innovations, iid, GARCH innovations (with two different parameter settings) and innovations following a stochastic volatility model (to be given below). The length 17

18 of the time series is 1000, with 1000 replications. 2. For each time series generated, estimate the EACF table and use the original inference procedure (under iid assumption) to identify the model order. 3. Of the 1000 replications, count the frequency (percentage) of each candidate models selected by the original EACF procedure 4. Repeat step 2 and 3 using the modified EACF procedure. The parameters for the ARMA(1, 1) mean equation is (0.8, 0.5) for the AR and MA coefficients respectively. We use four types of innovations: (i) iid; (ii) GARCH(1,1) with α = 0.1, β = 0.8; (iii) GARCH(1,1) with α = 0.5, β = 0.2; (iv) innovation follows a stochastic volatility with α 1 = 0.5 where the model is defined as ε t = σ t e t, (1 α 1 B) ln(σ 2 t ) = v t, (15) where e t s are iid N(0, 1), v t s are iid N(0, σv), 2 {e t } and {v t } are independent. The reason we simulate two settings of GARCH model is that the GARCH effects differ with different parameters. Min (2004) showed that Var( ˆρ) is significantly impacted if Cov(ε 2 0, ε2 q k+i )/E2 (ε 2 0 ) is large. For instance, if {ε t} is a GARCH(1,1) process as is the case in our simulation, Cov(ε 2 0, ε2 1 )/σ4 = 2α + (6α 2 (α + β/3))/(1 2α 2 (α + β) 2 ). Given α = 0.5 and β = 0.2 the ratio is 86. On the other hand, with α = 0.1 and β = 0.8, the ratio is The simulation results are reported in Table 7, showing the percentage of a candidate model being selected based on either the original inference or the modified inference of the EACF table. From Table 7, it is seen that the modified EACF procedure has similar performance as the original EACF for pure ARMA models, but outperforms it for ARMA+GARCH Models. We run a set of formal tests to see it more clearly. Since the same sets of simulated time series are used for original EACF and modified EACF procedure, we create a 2 2 contingency table for each innovation type and run chi-square test to compare identification power. The results are shown in Table 8 with the table and corresponding chi-square statistic and its p-value, for each innovation type. From Tables 7 and 8, we see that both procedures works equally well with iid innovations (large p-value), confirming that the two algorithms are almost 18

19 Table 7: EACF method comparison for a simulated ARMA(1,1) series with different type of innovations: X t = 0.8X t 1 + ε t + 0.5ε t 1 where innovation ε t s types are: (1) iid; (2) GARCH(1,1) with α = 0.1, β = 0.8; (3) GARCH(1,1) with α = 0.5, β = 0.2 (4) Stochastic Volatility with α 1 = 0.5. The time series length is 1000 while the replication is 1000 as well. The numbers (in %) represent the percentage which the order (in row) identified by the method (in column) for the simulated series. Each column would sum up to 100%. Innovation (i) Innovation (ii) Innovation (iii) Innovation (iv) Models original modified original modified original modified original modified white noise 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% MA(1) 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% 0.00% MA(2) 0.00% 0.00% 0.00% 0.00% 0.00% 0.20% 0.00% 0.00% MA(3) 0.00% 0.00% 0.00% 0.00% 0.00% 0.20% 0.00% 0.00% AR(1) 0.00% 0.00% 0.00% 0.00% 0.00% 0.30% 0.00% 0.00% ARMA(1,1) 81.40% 82.40% 76.90% 82.70% 66.20% 87.60% 70.10% 84.30% ARMA(1,2) 9.90% 8.50% 11.50% 7.60% 16.30% 4.40% 10.80% 7.40% ARMA(1,3) 4.10% 3.90% 5.30% 4.50% 7.50% 3.00% 6.10% 3.70% AR(2) 0.00% 0.00% 0.00% 0.00% 0.00% 0.10% 0.00% 0.00% ARMA(2,1) 0.20% 0.40% 0.40% 0.10% 0.80% 0.20% 0.40% 0.00% ARMA(2,2) 2.70% 3.10% 3.60% 3.30% 3.50% 2.10% 7.10% 3.00% ARMA(2,3) 0.80% 0.30% 0.80% 0.40% 0.70% 0.40% 1.50% 0.50% AR(3) 0.00% 0.00% 0.00% 0.10% 0.10% 0.20% 0.10% 0.00% ARMA(3,1) 0.00% 0.00% 0.10% 0.00% 0.20% 0.10% 0.00% 0.00% ARMA(2,2) 0.10% 0.20% 0.10% 0.20% 0.60% 0.20% 0.20% 0.10% ARMA(3,3) 0.50% 0.90% 0.70% 0.50% 1.90% 0.50% 2.30% 0.50% Others 0.30% 0.30% 0.60% 0.60% 2.20% 0.50% 1.40% 0.50% 19

20 Table 8: EACF procedure comparison by 2 2 contingency table and chi square test. Innovation (i) original modified ARMA(1,1) chi-square non ARMA(1,1) p-value Innovation (ii) original modified ARMA(1,1) chi-square non ARMA(1,1) p-value Innovation (iii) original modified ARMA(1,1) chi-square non ARMA(1,1) p-value 7.13E-30 Innovation (iv) original modified ARMA(1,1) chi-square non ARMA(1,1) p-value 3.78E-14 equivalent for iid error terms. With GARCH innovation but relatively weak dependence (innovation ii), the improvement by the modified EACF compared to original EACF is significant but not as strong as the stronger dependent GARCH innovation case (innovation iii). For the case of innovation following the Stochastic Volatility model, the effect is quite similar to that of GARCH situation. Hence the modified EACF procedure works for this type of innovations as well Autoregressive Integrated Model In this subsection, we study the ADF test for ARI process with dependent but uncorrelated innovations. As we discussed in the previous section, the ADF test continues to work in this case and we illustrate the conclusion by the following simulation. We use the following ARI(1, 1) model as the mean equation. { X t = φ X t 1 + ε t, ε t = g t ɛ t, g t = α 0 + αε 2 t 1 + βg t 1. The innovations used follow GARCH model with two parameter settings: α = 0.1, β = 0.8 and α = 0.5, β = 0.2. The results of 5% level test for 10,000 replications are reported in Table 9. It also include the results for an ARI(1, 0) to check the power of the test. 20

21 Table 9: Results of ADF Test: replication = 10 4 ; time series length = Model φ Innovations Percentage of rejection ARI(1,0) 0.95 iid % ARI(1,1) 0.95 iid 5.09% ARI(1,1) 0.95 α = 0.1, β = % ARI(1,1) 0.95 α = 0.5, β = % ARI(1,0) 0.8 iid % ARI(1,1) 0.8 iid 4.92% ARI(1,1) 0.8 α = 0.1, β = % ARI(1,1) 0.8 α = 0.5, β = % ARI(1,0) 0.5 iid % ARI(1,1) 0.5 iid 5.03% ARI(1,1) 0.5 α = 0.1, β = % ARI(1,1) 0.5 α = 0.5, β = % With a very large number of replication, we confirm that the ADF test maintains roughly the correct level under the null hypothesis, regardless the type of innovations, as our theorem indicates. 3.2 Real Data We analyze the following two real data sets utilizing our modified EACF procedure in this section General Motor Stock Price We analyze the log-return series of General Motor company (GM) for the period from January 2, 1996 to May 8, 2006, shown in Figure 1. Order determination schemes are then applied. The EACF significance tables with the original and modified estimated variance are shown in Table 10, where 0 indicates insignificant EACF coefficient at 5% level and 2 indicates a significant one. It is seen that the original inference procedure indicates an ARM A(2, 2) model while the modified EACF procedure indicates a white noise model. When ARMA(2,2)+GARCH(1,1) model in the form { X t = µ + φ 1 X t 1 + φ 2 X t 2 + ε t + ψ 1 ε t 1 + ψ 2 ε t 2, ε t = g t ɛ t, g t = α 0 + α 1 ε t 1 + β 1 g t 1, 21

22 is fitted to the data, all of the AR and MA coefficients are not significant, shown in Table 11. It shows that the modified EACF procedure provides more reliable model identification results. Although starting with a wrong model does not necessary result in an inaccurate final model, but one has to go through a more tedious and effort consuming estimation and model checking process. Figure 1: General Motor Stock Return Series. GE Daily Log return Series using the Closing Price Series logreturn Time Table 10: Comparison of EACF tables for GM return data. Original EACF Table Modified EACF Table [,1] [,2] [,3] [,4] [,5] [,6] [,7] [,1] [,2] [,3] [,4] [,5] [,6] [,7] [1,] [1,] [2,] [2,] [3,] [3,] [4,] [4,] [5,] [5,] [6,] [6,]

23 Table 11: The fitting results of GM data set using ARMA(2,2)+GARCH(1,1) model. Estimate Std. Error t value P r(> t ) Significance µ 2.37E E φ E E φ E E ψ E E ψ E E α E E α E E E-09 *** β E E <2.00E-16 *** Traffic Volume in a Large City Time series analysis is widely used in traffic control applications. Large amount of traffic information including speed and volume have been collected but accurate prediction remains a challenge. Here we analyze a traffic volume data collected in a major city in China. Each observation records the traffic volume (number of vehicles) passing a specific location is recorded every 3 minutes. The data set contains 8 weeks of data, with 480 observations each day. We use the first seven weeks for model identification and the last week to test the prediction results. Figure 2 shows four of the eight weeks of the time series. Clearly the time series possesses strong daily and weekly seasonality. To remove the seasonality and stabalize the variance, for each observation, we subtract from it the average of the volumes observed at the same time in the seven weeks, then divide it by the standard deviation of these seven volumes. Again, we use both the original EACF procedure and modified EACF procedure on the resulted series. Table 12 shows the result. According to the EACF table, the modified EACF procedure identifies an ARMA(1,1) while the original EACF procedure identifies an ARMA(1,4) model, if the significant entries are treated strictly. The estimation result for both models are shown in Table 13. The estimation shows that the last of the three extra MA terms in the ARMA(1,4)+GARCH(1,1) model is actually significant. The two models have very comparable BIC values (2.485 and 2.487). To further compare the two models, we obtain the multi-step prediction error using observation of the next 23

24 Figure 2: Big City Traffic Volume Data Set. week 1 week 2 Volumn Volumn Time Time week 3 week 4 Volumn Volumn Time Time Table 12: Comparison of EACF Tables for Traffic Data Set. Original EACF Table Modified EACF Table [,1] [,2] [,3] [,4] [,5] [,6] [,7] [,8] [,1] [,2] [,3] [,4] [,5] [,6] [,7] [,8] [1,] [1,] [2,] [2,] [3,] [3,] [4,] [4,] [5,] [5,] [6,] [6,]

25 Table 13: Estimation results for the traffic data ARMA(1,4)+GARCH(1,1) ARMA(1,1)+GARCH(1,1) Est Std. Err t value p value Est Std. Err t value p value ar1 9.90e e < 2e e e < 2e-16 ma1-8.81e e < 2e e e < 2e-16 ma2-2.58e e ma3-1.23e e ma4-2.79e e e-05 ω 5.44e e e e e e-10 α e e < 2e e e < 2e-16 β e e < 2e e e < 2e-16 week. Table 14 shows the Sum of Squares of Prediction Errors, defined as t (ˆx t(d) x t+d ) 2, where both the predictions and true observations are based on the transformed series. Here the d-step ahead prediction ˆx t (d) is an prediction of x t+d, made at time t with observations up to x t, and the model parameters estimated using the first seven weeks of data. It shows that the simpler model identified by the modified EACF procedure performs slightly better than the more complex one. Table 14: Sum of Squares of Prediction Errors: The numbers listed below are based on the summation of n = 3360 prediction error squares obtained from the eighth weeks transformed series. Sum of Squares of Prediction Error Model Prediction Steps d ARMA(1,4)+GARCH(1,1) ARMA(1,1)+GARCH(1,1) Appendix: Proof of the Theorems Proof of Theorem 1: Given conditions C.1 and C.2, Theorem 3 of Wu and Min (2005) implies that, for a given h, ( dist n(ˆγ(0) γ(0), ˆγ(1) γ(1),, ˆγ(h) γ(h)) N 0, (h+1) 1 Π (h+1) (h+1) where Π is the covariance matrix of random vector t= P 1(x t x t, x t 1 x t,, x t h x t ). 25 ),

26 By introducing a function H : R h+1 R h : H((x 0,, x h ) ) = (x 1 /x 0,, x h /x 0 ), with the argument of Delta method, we have the sample autocorrelation functions ˆρ(1),, ˆρ(h) follow asymptotic joint normal distribution, i.e. ( ) dist n(ˆρ(1) ρ(1),, ˆρ(h) ρ(h)) N 0, Σ, h 1 h h with Σ = DΠD. D R h (h+1) is the gradient of H( ) evaluated at (γ(0),, γ(h)), i.e. D = 1 γ(0) [ ρ h, I h ] where ρ h = (ρ 1,, ρ h ) and I h is the identity matrix of h h. This completes the proof for the first part of Theorem 1. For h > q, by the obvious fact ρ(h) = 0 for MA(q) process and the above expression for Σ, we have nˆρ(h) dist N(0, δ 2 h ), where δ 2 h = ξ 2 /γ 2 (0) and ξ = t= P 1x t h x t. Since {x t } follows MA(q) process x t = q i=1 ψ ix t i + ε t, we have ξ = ε 1 q i=0 ψ ix 1 h+i if h > q, therefore we have the following calculation for ξ 2, q ξ 2 = E(ε 1 ψ i x 1 h+i ) 2 i=0 = E[ε 2 1( κ i ε i ) 2 ] MA representation of {x t } = = i=0 κ i κ j E(ε 2 1ε i ε j ) = i,j=0 i=0 κ 2 i E(ε 2 1ε 2 i) i=0 κ 2 i E(ε 2 1)E(ε 2 i) + κ 2 i Cov(ε 2 1, ε 2 i). The first term above is i=0 i=0 i=0 i j terms vanish by assumption q E(ε 2 1) κ 2 i E(ε 2 i) = E(ε 2 1)E( κ 2 i ε 2 i) = E(ε 2 1)E( κ i ε i ) 2 = E(ε 2 1)E( ψ i x 1 h+i ) 2 = σ 2( q ψi 2 γ(0) + 2 i=0 = γ 2 (0) + 2 q γ 2 (k). k=1 q k=1 i=0 i=0 q ψ i ψ i+k γ(k) ) i=0 26

27 The last equality is due to the fact that σ 2( q ) i=0 ψ2 i = γ(0) and σ 2 ( q i=0 ψ ) iψ i+k = γ(k). Thus we have q q δh 2 (1 + 2 ρ 2 (k)) = ξ 2 /γ 2 (0) (1 + 2 ρ 2 (k)) = k=1 k=1 κ 2 i Cov(ε 2 1, ε 2 i)/γ 2 (0) i=0 Hence δ 2 h 1+2 q k=1 ρ2 (k) if i=0 κ2 i Cov(ε2 1, ε2 i )/γ2 (0) is nonnegative. This is true if all ACF of the ε 2 t series are non-negative. If all ACF are strictly positive, then δ 2 h > q k=1 ρ2 (k). In addition, when E(ε 2 n F n 1 ) = σ 2, we have the strict equality, which is the same as the iid case. To estimate the variance of ˆγ(h), h > q, note that Var(ˆγ(h)) = Var( 1 n h x t x t+h ) = 1 n h n n 2 Var( x t x t+h ) = MA(q) = n h n Symmetry = E(x 0 x h )=0 = t=1 t=1 1 ( n h n h n 2 Var(x t x t+h ) + 1 ( n 2 (n h)var(x 0 x h ) + 2 t=1 n h t=1 s=1,s t ) Cov(x t x t+h, x s x s+h ) q ) (n t)cov(x 0 x h, x t x t+h ) t=1 1 ( q Var(x 0 x h ) + 2 n t=1 1 Cov(x 0 x h, x d x d+h ) = 1 n n 1 n d q d q ) Cov(x 0 x h, x t x t+h ) d q ( ) E(x 0 x h x d x d+h ) E(x 0 x h )E(x d x d+h ) E(x 0 x h x d x d+h ) = 1 σ(d, h). (16) n d q Secondly, recall (4), we have ˆρ(h) = ˆγ(h) ˆγ(0). By Theorem 3 of Wu and Min (2005), [ ˆγ(0) ˆγ(h) ] ( [ dist γ(0) N γ(h) ] ), n 1 V. (17) 27

28 By Delta method (see Casella and Berger, 2002), we have for h > q Var(ˆρ(h)) = Var( ˆγ(h) ˆγ(0) ) ( γ(h) γ 2 (0), 1 γ(h)=0 = by (16) = ) ( [n 1 V ] γ(h) γ 2 (0), 1 ) T γ(0) γ(0) 1 γ 2 (0) n 1 V 22 Var(ˆγ(h)) γ 2 (0) 1 ( ) σ(0, h) + 2σ(1, h) + + 2σ(q, h). nγ 2 (0) Proof of Theorem 2: We adopt the notation X k R n k and ε R n 1 for any given k (p, p + s], x 0 x 1 x 1 k x X k = 1 x 0 x 2 k......, ε = x n 1 x n 2 x n k where {x t t = 1,, n} is the observed sequence and in the definition for X k we assume that x l = 0 for l 0. We first argue the asymptotic joint normal distribution for (X p+sε)/ n. In fact, with conditions C.1 and C.2, we can apply the Theorem 3 of Wu and Min (2005) respectively to new sequences of {x t j + ε t+1 } and {x t j ε t+1 } for any given j. Notice x t j ε t+1 = [(x t j +ε t+1 ) 2 (x t j ε t+1 ) 2 ]/4, we have P 1 (x t j ε t+1 ) <, which leads to the asymptotic normality of t=1 t=1 n x t j ε t+1 / n. Through Cramér-Wold device, we obtain asymptotic joint nor- mality of { n x t j ε t+1 / n j = 1,..., p + s}, i.e. (X p+sε)/ n. In particular, t=1 the mean vector of the asymptotic joint normal distribution is a vector of zeros since E(x t j ε t+1 ) = 0. On the other hand, lag-k PACF ˆφ k,k is obtained from the OLS regression of x t = φ 1 x t 1 + +φ p x t p + +φ k x t k where the true value of φ p+1,, φ k are all zero. For each k (p, p+s], n ˆφ k,k = [(X k X k/n) 1 ] (kth row) (X k ε)/ n is a linear transformation of the vector X p+sε/ n. Hence the vector n( ˆφ p+1,p+1,, ˆφ p+s,p+s ) also approaches a linear transformation of X p+sε/ n. Since (X p+sε)/ n follows asymptotic joint normal distribution with zero mean, it follows n( ˆφp+1,p+1,, ˆφ ( ) p+s,p+s ) dist N 0, Ξ. s 1 s s ε 1 ε 2. ε n, 28

29 This completes the proof for the first part of Theorem 2. To show (11), for simplicity of notations, lets denote k = p + l hereafter. Ξ (l,l) is related to the variance of asymptotic distribution of lag-k sample PACF ˆφ k,k. We first notice by the Theorem 3 of Wu and Min (2005), (X k ε)/ n has asymptotic normal distribution with covariance matrix Ω k = E(ξξ ) where ξ = t= (P 1 (ε t x t 1 ),, P 1 (ε t x t k )) = (ε 1 x 0, ε 1 x 1,, ε 1 x 1 k ). Since n ˆφ k,k = [(X k X k/n) 1 ] (kth row) (X k ε)/ n, we have Ξ (l,l) = [Γ 1 k Ω kγ 1 k ] (k,k), the (k, k) entry of matrix [Γ 1 k Ω kγ 1 k ] where Γ k := [γ(i j)] 1 i,j k. We will show Ω k σ 2 Γ k, i.e., Ω k σ 2 Γ k being a positive semidefinite matrix. To this end, consider any non-random vector C = (c 0, c 1,, c 1 k ) R k with a positive L 2 norm, C Ω k C = C [Cov(ε 1 x 0, ε 1 x 1,, ε 1 x 1 k ) ]C = E[ε 2 1(c 0 x 0 + c 1 x c 1 k x 1 k ) 2 ] = E[ε 2 1( κ i ε i ) 2 ] MA representation of {x t } = So Γ 1 k Ω kγ 1 k i=0 κ i κ j E(ε 2 1ε i ε j ) = i,j=0 κ 2 i E(ε 2 1)E(ε 2 i) i=0 κ 2 i E(ε 2 1ε 2 i) i=0 i j terms vanish by assumption ACF of {ε 2 t } are nonnegative by assumption = σ 2 κ 2 i E(ε 2 i) = σ 2 E[(c 0 x 0 + c 1 x c 1 k x 1 k ) 2 ] i=0 = σ 2 C Γ k C > 0 Γ 1 k σ2 Γ k Γ 1 k = σ 2 Γ 1 k. It is known that for AR(p) process the (k, k) entry of Γ 1 k is 1/σ 2 for k > p. The result follows in view of [Γ 1 k Ω kγ 1 k ] (k,k) σ 2 [Γ 1 k ] (k,k) = 1. If all ACFs are positive, then the inequality in above derivations holds strictly since Ω k > σ 2 Γ k. Let S n = n t=1 ε t, then Sn 2 2 = E( n t=1 ε t) 2 = E( n t=1 ε2 t ). Denote Vn 2 S n 2 2 = n t=1 ε2 t. For k = 0,, n, set { Sk V n, t = k n ; W n (t) = S k V n + (nt k) ε k+1 V n, k n t k+1 n. = 29

30 Lemma 3 For time series {X t } following (12), let θ i be the coefficients at lag i of the MA infinity representation of {X t }, Θ n = n i=0 θ i, Bn 2 = n 1 i=0 Θ2 i. Assume that the error series {ε t } satisfies condition C.3, i=1 (Θ n+i Θ i ) 2 = o(bn) 2 and B n, then W n dist W. Remark 6 The condition i=1 (Θ n+i Θ i ) 2 = o(bn) 2 is related to the mean equation s coefficients and is automatically satisfied for stationary and invertible time series. Proof of Lemma 3: See Theorem 1 of Wu and Min (2005). Proof of Theorem 3: Recall that S n = n t=1 ε t and Vn 2 = S n 2 2 = n t=1 ε2 t. Also recall that ˆρ n, the least square estimator of ρ, has the following form ρ n = n t=p+1 X t 1(X t p 1 k=1 âk X t k ). n t=p+1 X2 t 1 Under null hypothesis where ρ = 1, since no MA part is involved in X t, from multiple regression property â k a k, we immediately have T n = n( ρ n 1) = n n t=p+1 X t 1(X t X t 1 p 1 k=1 âk X t k ) n t=p+1 X2 t 1 = 1 n t=p+1 n X t 1ε t + o p (1). 1 n n 2 t=p+1 X2 t 1 Denote c = (1 p 1 k=1 a k) 1, and use the same polynomial decomposition given by Theorem of Fuller (1995). Although the conditions on innovations are different, they do not impact the argument for this specific part. Using the same argument which still hold for the our assumptions, we get 1 [ n 2 1 [ n n t=p+1 n t=p+1 X 2 t 1 c 2 X t 1 ε t c n t=p+1 n t=p+1 S 2 t 1] = op (1); S t 1 ε t ] = op (1). 30

STAT 443 Final Exam Review. 1 Basic Definitions. 2 Statistical Tests. L A TEXer: W. Kong

STAT 443 Final Exam Review. 1 Basic Definitions. 2 Statistical Tests. L A TEXer: W. Kong STAT 443 Final Exam Review L A TEXer: W Kong 1 Basic Definitions Definition 11 The time series {X t } with E[X 2 t ] < is said to be weakly stationary if: 1 µ X (t) = E[X t ] is independent of t 2 γ X

More information

FE570 Financial Markets and Trading. Stevens Institute of Technology

FE570 Financial Markets and Trading. Stevens Institute of Technology FE570 Financial Markets and Trading Lecture 5. Linear Time Series Analysis and Its Applications (Ref. Joel Hasbrouck - Empirical Market Microstructure ) Steve Yang Stevens Institute of Technology 9/25/2012

More information

Ch 6. Model Specification. Time Series Analysis

Ch 6. Model Specification. Time Series Analysis We start to build ARIMA(p,d,q) models. The subjects include: 1 how to determine p, d, q for a given series (Chapter 6); 2 how to estimate the parameters (φ s and θ s) of a specific ARIMA(p,d,q) model (Chapter

More information

Review Session: Econometrics - CLEFIN (20192)

Review Session: Econometrics - CLEFIN (20192) Review Session: Econometrics - CLEFIN (20192) Part II: Univariate time series analysis Daniele Bianchi March 20, 2013 Fundamentals Stationarity A time series is a sequence of random variables x t, t =

More information

AR, MA and ARMA models

AR, MA and ARMA models AR, MA and AR by Hedibert Lopes P Based on Tsay s Analysis of Financial Time Series (3rd edition) P 1 Stationarity 2 3 4 5 6 7 P 8 9 10 11 Outline P Linear Time Series Analysis and Its Applications For

More information

Applied time-series analysis

Applied time-series analysis Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna October 18, 2011 Outline Introduction and overview Econometric Time-Series Analysis In principle,

More information

Multivariate Time Series

Multivariate Time Series Multivariate Time Series Notation: I do not use boldface (or anything else) to distinguish vectors from scalars. Tsay (and many other writers) do. I denote a multivariate stochastic process in the form

More information

Lecture 7: Model Building Bus 41910, Time Series Analysis, Mr. R. Tsay

Lecture 7: Model Building Bus 41910, Time Series Analysis, Mr. R. Tsay Lecture 7: Model Building Bus 41910, Time Series Analysis, Mr R Tsay An effective procedure for building empirical time series models is the Box-Jenkins approach, which consists of three stages: model

More information

STAT Financial Time Series

STAT Financial Time Series STAT 6104 - Financial Time Series Chapter 4 - Estimation in the time Domain Chun Yip Yau (CUHK) STAT 6104:Financial Time Series 1 / 46 Agenda 1 Introduction 2 Moment Estimates 3 Autoregressive Models (AR

More information

Econometric Forecasting

Econometric Forecasting Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna October 1, 2014 Outline Introduction Model-free extrapolation Univariate time-series models Trend

More information

Empirical Market Microstructure Analysis (EMMA)

Empirical Market Microstructure Analysis (EMMA) Empirical Market Microstructure Analysis (EMMA) Lecture 3: Statistical Building Blocks and Econometric Basics Prof. Dr. Michael Stein michael.stein@vwl.uni-freiburg.de Albert-Ludwigs-University of Freiburg

More information

Time Series Forecasting: A Tool for Out - Sample Model Selection and Evaluation

Time Series Forecasting: A Tool for Out - Sample Model Selection and Evaluation AMERICAN JOURNAL OF SCIENTIFIC AND INDUSTRIAL RESEARCH 214, Science Huβ, http://www.scihub.org/ajsir ISSN: 2153-649X, doi:1.5251/ajsir.214.5.6.185.194 Time Series Forecasting: A Tool for Out - Sample Model

More information

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY Time Series Analysis James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY & Contents PREFACE xiii 1 1.1. 1.2. Difference Equations First-Order Difference Equations 1 /?th-order Difference

More information

Nonlinear time series

Nonlinear time series Based on the book by Fan/Yao: Nonlinear Time Series Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna October 27, 2009 Outline Characteristics of

More information

Autoregressive Moving Average (ARMA) Models and their Practical Applications

Autoregressive Moving Average (ARMA) Models and their Practical Applications Autoregressive Moving Average (ARMA) Models and their Practical Applications Massimo Guidolin February 2018 1 Essential Concepts in Time Series Analysis 1.1 Time Series and Their Properties Time series:

More information

Chapter 12: An introduction to Time Series Analysis. Chapter 12: An introduction to Time Series Analysis

Chapter 12: An introduction to Time Series Analysis. Chapter 12: An introduction to Time Series Analysis Chapter 12: An introduction to Time Series Analysis Introduction In this chapter, we will discuss forecasting with single-series (univariate) Box-Jenkins models. The common name of the models is Auto-Regressive

More information

Time Series I Time Domain Methods

Time Series I Time Domain Methods Astrostatistics Summer School Penn State University University Park, PA 16802 May 21, 2007 Overview Filtering and the Likelihood Function Time series is the study of data consisting of a sequence of DEPENDENT

More information

Dynamic Time Series Regression: A Panacea for Spurious Correlations

Dynamic Time Series Regression: A Panacea for Spurious Correlations International Journal of Scientific and Research Publications, Volume 6, Issue 10, October 2016 337 Dynamic Time Series Regression: A Panacea for Spurious Correlations Emmanuel Alphonsus Akpan *, Imoh

More information

Chapter 6: Model Specification for Time Series

Chapter 6: Model Specification for Time Series Chapter 6: Model Specification for Time Series The ARIMA(p, d, q) class of models as a broad class can describe many real time series. Model specification for ARIMA(p, d, q) models involves 1. Choosing

More information

Topic 4 Unit Roots. Gerald P. Dwyer. February Clemson University

Topic 4 Unit Roots. Gerald P. Dwyer. February Clemson University Topic 4 Unit Roots Gerald P. Dwyer Clemson University February 2016 Outline 1 Unit Roots Introduction Trend and Difference Stationary Autocorrelations of Series That Have Deterministic or Stochastic Trends

More information

Prof. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis

Prof. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis Introduction to Time Series Analysis 1 Contents: I. Basics of Time Series Analysis... 4 I.1 Stationarity... 5 I.2 Autocorrelation Function... 9 I.3 Partial Autocorrelation Function (PACF)... 14 I.4 Transformation

More information

at least 50 and preferably 100 observations should be available to build a proper model

at least 50 and preferably 100 observations should be available to build a proper model III Box-Jenkins Methods 1. Pros and Cons of ARIMA Forecasting a) need for data at least 50 and preferably 100 observations should be available to build a proper model used most frequently for hourly or

More information

Time Series 2. Robert Almgren. Sept. 21, 2009

Time Series 2. Robert Almgren. Sept. 21, 2009 Time Series 2 Robert Almgren Sept. 21, 2009 This week we will talk about linear time series models: AR, MA, ARMA, ARIMA, etc. First we will talk about theory and after we will talk about fitting the models

More information

Akaike criterion: Kullback-Leibler discrepancy

Akaike criterion: Kullback-Leibler discrepancy Model choice. Akaike s criterion Akaike criterion: Kullback-Leibler discrepancy Given a family of probability densities {f ( ; ψ), ψ Ψ}, Kullback-Leibler s index of f ( ; ψ) relative to f ( ; θ) is (ψ

More information

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY Time Series Analysis James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY PREFACE xiii 1 Difference Equations 1.1. First-Order Difference Equations 1 1.2. pth-order Difference Equations 7

More information

Lecture 2: Univariate Time Series

Lecture 2: Univariate Time Series Lecture 2: Univariate Time Series Analysis: Conditional and Unconditional Densities, Stationarity, ARMA Processes Prof. Massimo Guidolin 20192 Financial Econometrics Spring/Winter 2017 Overview Motivation:

More information

6. The econometrics of Financial Markets: Empirical Analysis of Financial Time Series. MA6622, Ernesto Mordecki, CityU, HK, 2006.

6. The econometrics of Financial Markets: Empirical Analysis of Financial Time Series. MA6622, Ernesto Mordecki, CityU, HK, 2006. 6. The econometrics of Financial Markets: Empirical Analysis of Financial Time Series MA6622, Ernesto Mordecki, CityU, HK, 2006. References for Lecture 5: Quantitative Risk Management. A. McNeil, R. Frey,

More information

Econ 423 Lecture Notes: Additional Topics in Time Series 1

Econ 423 Lecture Notes: Additional Topics in Time Series 1 Econ 423 Lecture Notes: Additional Topics in Time Series 1 John C. Chao April 25, 2017 1 These notes are based in large part on Chapter 16 of Stock and Watson (2011). They are for instructional purposes

More information

Advanced Econometrics

Advanced Econometrics Advanced Econometrics Marco Sunder Nov 04 2010 Marco Sunder Advanced Econometrics 1/ 25 Contents 1 2 3 Marco Sunder Advanced Econometrics 2/ 25 Music Marco Sunder Advanced Econometrics 3/ 25 Music Marco

More information

Test for Parameter Change in ARIMA Models

Test for Parameter Change in ARIMA Models Test for Parameter Change in ARIMA Models Sangyeol Lee 1 Siyun Park 2 Koichi Maekawa 3 and Ken-ichi Kawai 4 Abstract In this paper we consider the problem of testing for parameter changes in ARIMA models

More information

APPLIED ECONOMETRIC TIME SERIES 4TH EDITION

APPLIED ECONOMETRIC TIME SERIES 4TH EDITION APPLIED ECONOMETRIC TIME SERIES 4TH EDITION Chapter 2: STATIONARY TIME-SERIES MODELS WALTER ENDERS, UNIVERSITY OF ALABAMA Copyright 2015 John Wiley & Sons, Inc. Section 1 STOCHASTIC DIFFERENCE EQUATION

More information

Stationary Stochastic Time Series Models

Stationary Stochastic Time Series Models Stationary Stochastic Time Series Models When modeling time series it is useful to regard an observed time series, (x 1,x,..., x n ), as the realisation of a stochastic process. In general a stochastic

More information

GARCH Models. Eduardo Rossi University of Pavia. December Rossi GARCH Financial Econometrics / 50

GARCH Models. Eduardo Rossi University of Pavia. December Rossi GARCH Financial Econometrics / 50 GARCH Models Eduardo Rossi University of Pavia December 013 Rossi GARCH Financial Econometrics - 013 1 / 50 Outline 1 Stylized Facts ARCH model: definition 3 GARCH model 4 EGARCH 5 Asymmetric Models 6

More information

Forecasting. Simon Shaw 2005/06 Semester II

Forecasting. Simon Shaw 2005/06 Semester II Forecasting Simon Shaw s.c.shaw@maths.bath.ac.uk 2005/06 Semester II 1 Introduction A critical aspect of managing any business is planning for the future. events is called forecasting. Predicting future

More information

STAT 248: EDA & Stationarity Handout 3

STAT 248: EDA & Stationarity Handout 3 STAT 248: EDA & Stationarity Handout 3 GSI: Gido van de Ven September 17th, 2010 1 Introduction Today s section we will deal with the following topics: the mean function, the auto- and crosscovariance

More information

Arma-Arch Modeling Of The Returns Of First Bank Of Nigeria

Arma-Arch Modeling Of The Returns Of First Bank Of Nigeria Arma-Arch Modeling Of The Returns Of First Bank Of Nigeria Emmanuel Alphonsus Akpan Imoh Udo Moffat Department of Mathematics and Statistics University of Uyo, Nigeria Ntiedo Bassey Ekpo Department of

More information

Multivariate Time Series: VAR(p) Processes and Models

Multivariate Time Series: VAR(p) Processes and Models Multivariate Time Series: VAR(p) Processes and Models A VAR(p) model, for p > 0 is X t = φ 0 + Φ 1 X t 1 + + Φ p X t p + A t, where X t, φ 0, and X t i are k-vectors, Φ 1,..., Φ p are k k matrices, with

More information

3. ARMA Modeling. Now: Important class of stationary processes

3. ARMA Modeling. Now: Important class of stationary processes 3. ARMA Modeling Now: Important class of stationary processes Definition 3.1: (ARMA(p, q) process) Let {ɛ t } t Z WN(0, σ 2 ) be a white noise process. The process {X t } t Z is called AutoRegressive-Moving-Average

More information

TIME SERIES ANALYSIS AND FORECASTING USING THE STATISTICAL MODEL ARIMA

TIME SERIES ANALYSIS AND FORECASTING USING THE STATISTICAL MODEL ARIMA CHAPTER 6 TIME SERIES ANALYSIS AND FORECASTING USING THE STATISTICAL MODEL ARIMA 6.1. Introduction A time series is a sequence of observations ordered in time. A basic assumption in the time series analysis

More information

Parameter estimation: ACVF of AR processes

Parameter estimation: ACVF of AR processes Parameter estimation: ACVF of AR processes Yule-Walker s for AR processes: a method of moments, i.e. µ = x and choose parameters so that γ(h) = ˆγ(h) (for h small ). 12 novembre 2013 1 / 8 Parameter estimation:

More information

Lesson 13: Box-Jenkins Modeling Strategy for building ARMA models

Lesson 13: Box-Jenkins Modeling Strategy for building ARMA models Lesson 13: Box-Jenkins Modeling Strategy for building ARMA models Facoltà di Economia Università dell Aquila umberto.triacca@gmail.com Introduction In this lesson we present a method to construct an ARMA(p,

More information

The autocorrelation and autocovariance functions - helpful tools in the modelling problem

The autocorrelation and autocovariance functions - helpful tools in the modelling problem The autocorrelation and autocovariance functions - helpful tools in the modelling problem J. Nowicka-Zagrajek A. Wy lomańska Institute of Mathematics and Computer Science Wroc law University of Technology,

More information

STAT STOCHASTIC PROCESSES. Contents

STAT STOCHASTIC PROCESSES. Contents STAT 3911 - STOCHASTIC PROCESSES ANDREW TULLOCH Contents 1. Stochastic Processes 2 2. Classification of states 2 3. Limit theorems for Markov chains 4 4. First step analysis 5 5. Branching processes 5

More information

Some Time-Series Models

Some Time-Series Models Some Time-Series Models Outline 1. Stochastic processes and their properties 2. Stationary processes 3. Some properties of the autocorrelation function 4. Some useful models Purely random processes, random

More information

Time Series Analysis -- An Introduction -- AMS 586

Time Series Analysis -- An Introduction -- AMS 586 Time Series Analysis -- An Introduction -- AMS 586 1 Objectives of time series analysis Data description Data interpretation Modeling Control Prediction & Forecasting 2 Time-Series Data Numerical data

More information

Stat 248 Lab 2: Stationarity, More EDA, Basic TS Models

Stat 248 Lab 2: Stationarity, More EDA, Basic TS Models Stat 248 Lab 2: Stationarity, More EDA, Basic TS Models Tessa L. Childers-Day February 8, 2013 1 Introduction Today s section will deal with topics such as: the mean function, the auto- and cross-covariance

More information

A TIME SERIES PARADOX: UNIT ROOT TESTS PERFORM POORLY WHEN DATA ARE COINTEGRATED

A TIME SERIES PARADOX: UNIT ROOT TESTS PERFORM POORLY WHEN DATA ARE COINTEGRATED A TIME SERIES PARADOX: UNIT ROOT TESTS PERFORM POORLY WHEN DATA ARE COINTEGRATED by W. Robert Reed Department of Economics and Finance University of Canterbury, New Zealand Email: bob.reed@canterbury.ac.nz

More information

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley Time Series Models and Inference James L. Powell Department of Economics University of California, Berkeley Overview In contrast to the classical linear regression model, in which the components of the

More information

Introduction to Stochastic processes

Introduction to Stochastic processes Università di Pavia Introduction to Stochastic processes Eduardo Rossi Stochastic Process Stochastic Process: A stochastic process is an ordered sequence of random variables defined on a probability space

More information

Time Series Outlier Detection

Time Series Outlier Detection Time Series Outlier Detection Tingyi Zhu July 28, 2016 Tingyi Zhu Time Series Outlier Detection July 28, 2016 1 / 42 Outline Time Series Basics Outliers Detection in Single Time Series Outlier Series Detection

More information

Stochastic Modelling Solutions to Exercises on Time Series

Stochastic Modelling Solutions to Exercises on Time Series Stochastic Modelling Solutions to Exercises on Time Series Dr. Iqbal Owadally March 3, 2003 Solutions to Elementary Problems Q1. (i) (1 0.5B)X t = Z t. The characteristic equation 1 0.5z = 0 does not have

More information

13. Time Series Analysis: Asymptotics Weakly Dependent and Random Walk Process. Strict Exogeneity

13. Time Series Analysis: Asymptotics Weakly Dependent and Random Walk Process. Strict Exogeneity Outline: Further Issues in Using OLS with Time Series Data 13. Time Series Analysis: Asymptotics Weakly Dependent and Random Walk Process I. Stationary and Weakly Dependent Time Series III. Highly Persistent

More information

Introduction to Time Series Analysis. Lecture 11.

Introduction to Time Series Analysis. Lecture 11. Introduction to Time Series Analysis. Lecture 11. Peter Bartlett 1. Review: Time series modelling and forecasting 2. Parameter estimation 3. Maximum likelihood estimator 4. Yule-Walker estimation 5. Yule-Walker

More information

Econometría 2: Análisis de series de Tiempo

Econometría 2: Análisis de series de Tiempo Econometría 2: Análisis de series de Tiempo Karoll GOMEZ kgomezp@unal.edu.co http://karollgomez.wordpress.com Segundo semestre 2016 IX. Vector Time Series Models VARMA Models A. 1. Motivation: The vector

More information

{ } Stochastic processes. Models for time series. Specification of a process. Specification of a process. , X t3. ,...X tn }

{ } Stochastic processes. Models for time series. Specification of a process. Specification of a process. , X t3. ,...X tn } Stochastic processes Time series are an example of a stochastic or random process Models for time series A stochastic process is 'a statistical phenomenon that evolves in time according to probabilistic

More information

Time Series 4. Robert Almgren. Oct. 5, 2009

Time Series 4. Robert Almgren. Oct. 5, 2009 Time Series 4 Robert Almgren Oct. 5, 2009 1 Nonstationarity How should you model a process that has drift? ARMA models are intrinsically stationary, that is, they are mean-reverting: when the value of

More information

7. Integrated Processes

7. Integrated Processes 7. Integrated Processes Up to now: Analysis of stationary processes (stationary ARMA(p, q) processes) Problem: Many economic time series exhibit non-stationary patterns over time 226 Example: We consider

More information

Basics: Definitions and Notation. Stationarity. A More Formal Definition

Basics: Definitions and Notation. Stationarity. A More Formal Definition Basics: Definitions and Notation A Univariate is a sequence of measurements of the same variable collected over (usually regular intervals of) time. Usual assumption in many time series techniques is that

More information

1 Time Series Concepts and Challenges

1 Time Series Concepts and Challenges Forecasting Time Series Data Notes from Rebecca Sela Stern Business School Spring, 2004 1 Time Series Concepts and Challenges The linear regression model (and most other models) assume that observations

More information

Multivariate Time Series Analysis and Its Applications [Tsay (2005), chapter 8]

Multivariate Time Series Analysis and Its Applications [Tsay (2005), chapter 8] 1 Multivariate Time Series Analysis and Its Applications [Tsay (2005), chapter 8] Insights: Price movements in one market can spread easily and instantly to another market [economic globalization and internet

More information

Econ 623 Econometrics II Topic 2: Stationary Time Series

Econ 623 Econometrics II Topic 2: Stationary Time Series 1 Introduction Econ 623 Econometrics II Topic 2: Stationary Time Series In the regression model we can model the error term as an autoregression AR(1) process. That is, we can use the past value of the

More information

FORECASTING SUGARCANE PRODUCTION IN INDIA WITH ARIMA MODEL

FORECASTING SUGARCANE PRODUCTION IN INDIA WITH ARIMA MODEL FORECASTING SUGARCANE PRODUCTION IN INDIA WITH ARIMA MODEL B. N. MANDAL Abstract: Yearly sugarcane production data for the period of - to - of India were analyzed by time-series methods. Autocorrelation

More information

STOR 356: Summary Course Notes

STOR 356: Summary Course Notes STOR 356: Summary Course Notes Richard L. Smith Department of Statistics and Operations Research University of North Carolina Chapel Hill, NC 7599-360 rls@email.unc.edu February 19, 008 Course text: Introduction

More information

University of Oxford. Statistical Methods Autocorrelation. Identification and Estimation

University of Oxford. Statistical Methods Autocorrelation. Identification and Estimation University of Oxford Statistical Methods Autocorrelation Identification and Estimation Dr. Órlaith Burke Michaelmas Term, 2011 Department of Statistics, 1 South Parks Road, Oxford OX1 3TG Contents 1 Model

More information

Oil price volatility in the Philippines using generalized autoregressive conditional heteroscedasticity

Oil price volatility in the Philippines using generalized autoregressive conditional heteroscedasticity Oil price volatility in the Philippines using generalized autoregressive conditional heteroscedasticity Carl Ceasar F. Talungon University of Southern Mindanao, Cotabato Province, Philippines Email: carlceasar04@gmail.com

More information

7. Integrated Processes

7. Integrated Processes 7. Integrated Processes Up to now: Analysis of stationary processes (stationary ARMA(p, q) processes) Problem: Many economic time series exhibit non-stationary patterns over time 226 Example: We consider

More information

Lecture 1: Fundamental concepts in Time Series Analysis (part 2)

Lecture 1: Fundamental concepts in Time Series Analysis (part 2) Lecture 1: Fundamental concepts in Time Series Analysis (part 2) Florian Pelgrin University of Lausanne, École des HEC Department of mathematics (IMEA-Nice) Sept. 2011 - Jan. 2012 Florian Pelgrin (HEC)

More information

EASTERN MEDITERRANEAN UNIVERSITY ECON 604, FALL 2007 DEPARTMENT OF ECONOMICS MEHMET BALCILAR ARIMA MODELS: IDENTIFICATION

EASTERN MEDITERRANEAN UNIVERSITY ECON 604, FALL 2007 DEPARTMENT OF ECONOMICS MEHMET BALCILAR ARIMA MODELS: IDENTIFICATION ARIMA MODELS: IDENTIFICATION A. Autocorrelations and Partial Autocorrelations 1. Summary of What We Know So Far: a) Series y t is to be modeled by Box-Jenkins methods. The first step was to convert y t

More information

Vector autoregressions, VAR

Vector autoregressions, VAR 1 / 45 Vector autoregressions, VAR Chapter 2 Financial Econometrics Michael Hauser WS17/18 2 / 45 Content Cross-correlations VAR model in standard/reduced form Properties of VAR(1), VAR(p) Structural VAR,

More information

Econometrics I: Univariate Time Series Econometrics (1)

Econometrics I: Univariate Time Series Econometrics (1) Econometrics I: Dipartimento di Economia Politica e Metodi Quantitativi University of Pavia Overview of the Lecture 1 st EViews Session VI: Some Theoretical Premises 2 Overview of the Lecture 1 st EViews

More information

Contents. 1 Time Series Analysis Introduction Stationary Processes State Space Modesl Stationary Processes 8

Contents. 1 Time Series Analysis Introduction Stationary Processes State Space Modesl Stationary Processes 8 A N D R E W T U L L O C H T I M E S E R I E S A N D M O N T E C A R L O I N F E R E N C E T R I N I T Y C O L L E G E T H E U N I V E R S I T Y O F C A M B R I D G E Contents 1 Time Series Analysis 5

More information

7. Forecasting with ARIMA models

7. Forecasting with ARIMA models 7. Forecasting with ARIMA models 309 Outline: Introduction The prediction equation of an ARIMA model Interpreting the predictions Variance of the predictions Forecast updating Measuring predictability

More information

Note: The primary reference for these notes is Enders (2004). An alternative and more technical treatment can be found in Hamilton (1994).

Note: The primary reference for these notes is Enders (2004). An alternative and more technical treatment can be found in Hamilton (1994). Chapter 4 Analysis of a Single Time Series Note: The primary reference for these notes is Enders (4). An alternative and more technical treatment can be found in Hamilton (994). Most data used in financial

More information

Introduction to ARMA and GARCH processes

Introduction to ARMA and GARCH processes Introduction to ARMA and GARCH processes Fulvio Corsi SNS Pisa 3 March 2010 Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 1 / 24 Stationarity Strict stationarity: (X 1,

More information

A SARIMAX coupled modelling applied to individual load curves intraday forecasting

A SARIMAX coupled modelling applied to individual load curves intraday forecasting A SARIMAX coupled modelling applied to individual load curves intraday forecasting Frédéric Proïa Workshop EDF Institut Henri Poincaré - Paris 05 avril 2012 INRIA Bordeaux Sud-Ouest Institut de Mathématiques

More information

Heteroskedasticity; Step Changes; VARMA models; Likelihood ratio test statistic; Cusum statistic.

Heteroskedasticity; Step Changes; VARMA models; Likelihood ratio test statistic; Cusum statistic. 47 3!,57 Statistics and Econometrics Series 5 Febrary 24 Departamento de Estadística y Econometría Universidad Carlos III de Madrid Calle Madrid, 126 2893 Getafe (Spain) Fax (34) 91 624-98-49 VARIANCE

More information

Stochastic Processes

Stochastic Processes Stochastic Processes Stochastic Process Non Formal Definition: Non formal: A stochastic process (random process) is the opposite of a deterministic process such as one defined by a differential equation.

More information

Forecasting using R. Rob J Hyndman. 2.4 Non-seasonal ARIMA models. Forecasting using R 1

Forecasting using R. Rob J Hyndman. 2.4 Non-seasonal ARIMA models. Forecasting using R 1 Forecasting using R Rob J Hyndman 2.4 Non-seasonal ARIMA models Forecasting using R 1 Outline 1 Autoregressive models 2 Moving average models 3 Non-seasonal ARIMA models 4 Partial autocorrelations 5 Estimation

More information

STAT 443 (Winter ) Forecasting

STAT 443 (Winter ) Forecasting Winter 2014 TABLE OF CONTENTS STAT 443 (Winter 2014-1141) Forecasting Prof R Ramezan University of Waterloo L A TEXer: W KONG http://wwkonggithubio Last Revision: September 3, 2014 Table of Contents 1

More information

Unit Root and Cointegration

Unit Root and Cointegration Unit Root and Cointegration Carlos Hurtado Department of Economics University of Illinois at Urbana-Champaign hrtdmrt@illinois.edu Oct 7th, 016 C. Hurtado (UIUC - Economics) Applied Econometrics On the

More information

Lecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications

Lecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications Lecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications Prof. Massimo Guidolin 20192 Financial Econometrics Winter/Spring 2018 Overview Moving average processes Autoregressive

More information

1 Linear Difference Equations

1 Linear Difference Equations ARMA Handout Jialin Yu 1 Linear Difference Equations First order systems Let {ε t } t=1 denote an input sequence and {y t} t=1 sequence generated by denote an output y t = φy t 1 + ε t t = 1, 2,... with

More information

Part 1. Multiple Choice (50 questions, 1 point each) Part 2. Problems/Short Answer (10 questions, 5 points each)

Part 1. Multiple Choice (50 questions, 1 point each) Part 2. Problems/Short Answer (10 questions, 5 points each) GROUND RULES: This exam contains two parts: Part 1. Multiple Choice (50 questions, 1 point each) Part 2. Problems/Short Answer (10 questions, 5 points each) The maximum number of points on this exam is

More information

Part II. Time Series

Part II. Time Series Part II Time Series 12 Introduction This Part is mainly a summary of the book of Brockwell and Davis (2002). Additionally the textbook Shumway and Stoffer (2010) can be recommended. 1 Our purpose is to

More information

γ 0 = Var(X i ) = Var(φ 1 X i 1 +W i ) = φ 2 1γ 0 +σ 2, which implies that we must have φ 1 < 1, and γ 0 = σ2 . 1 φ 2 1 We may also calculate for j 1

γ 0 = Var(X i ) = Var(φ 1 X i 1 +W i ) = φ 2 1γ 0 +σ 2, which implies that we must have φ 1 < 1, and γ 0 = σ2 . 1 φ 2 1 We may also calculate for j 1 4.2 Autoregressive (AR) Moving average models are causal linear processes by definition. There is another class of models, based on a recursive formulation similar to the exponentially weighted moving

More information

ECON 616: Lecture 1: Time Series Basics

ECON 616: Lecture 1: Time Series Basics ECON 616: Lecture 1: Time Series Basics ED HERBST August 30, 2017 References Overview: Chapters 1-3 from Hamilton (1994). Technical Details: Chapters 2-3 from Brockwell and Davis (1987). Intuition: Chapters

More information

Univariate ARIMA Models

Univariate ARIMA Models Univariate ARIMA Models ARIMA Model Building Steps: Identification: Using graphs, statistics, ACFs and PACFs, transformations, etc. to achieve stationary and tentatively identify patterns and model components.

More information

10. Time series regression and forecasting

10. Time series regression and forecasting 10. Time series regression and forecasting Key feature of this section: Analysis of data on a single entity observed at multiple points in time (time series data) Typical research questions: What is the

More information

Problem Set 2: Box-Jenkins methodology

Problem Set 2: Box-Jenkins methodology Problem Set : Box-Jenkins methodology 1) For an AR1) process we have: γ0) = σ ε 1 φ σ ε γ0) = 1 φ Hence, For a MA1) process, p lim R = φ γ0) = 1 + θ )σ ε σ ε 1 = γ0) 1 + θ Therefore, p lim R = 1 1 1 +

More information

Model selection using penalty function criteria

Model selection using penalty function criteria Model selection using penalty function criteria Laimonis Kavalieris University of Otago Dunedin, New Zealand Econometrics, Time Series Analysis, and Systems Theory Wien, June 18 20 Outline Classes of models.

More information

Covariances of ARMA Processes

Covariances of ARMA Processes Statistics 910, #10 1 Overview Covariances of ARMA Processes 1. Review ARMA models: causality and invertibility 2. AR covariance functions 3. MA and ARMA covariance functions 4. Partial autocorrelation

More information

NANYANG TECHNOLOGICAL UNIVERSITY SEMESTER II EXAMINATION MAS451/MTH451 Time Series Analysis TIME ALLOWED: 2 HOURS

NANYANG TECHNOLOGICAL UNIVERSITY SEMESTER II EXAMINATION MAS451/MTH451 Time Series Analysis TIME ALLOWED: 2 HOURS NANYANG TECHNOLOGICAL UNIVERSITY SEMESTER II EXAMINATION 2012-2013 MAS451/MTH451 Time Series Analysis May 2013 TIME ALLOWED: 2 HOURS INSTRUCTIONS TO CANDIDATES 1. This examination paper contains FOUR (4)

More information

Estimation and application of best ARIMA model for forecasting the uranium price.

Estimation and application of best ARIMA model for forecasting the uranium price. Estimation and application of best ARIMA model for forecasting the uranium price. Medeu Amangeldi May 13, 2018 Capstone Project Superviser: Dongming Wei Second reader: Zhenisbek Assylbekov Abstract This

More information

Part III Example Sheet 1 - Solutions YC/Lent 2015 Comments and corrections should be ed to

Part III Example Sheet 1 - Solutions YC/Lent 2015 Comments and corrections should be  ed to TIME SERIES Part III Example Sheet 1 - Solutions YC/Lent 2015 Comments and corrections should be emailed to Y.Chen@statslab.cam.ac.uk. 1. Let {X t } be a weakly stationary process with mean zero and let

More information

Non-Stationary Time Series and Unit Root Testing

Non-Stationary Time Series and Unit Root Testing Econometrics II Non-Stationary Time Series and Unit Root Testing Morten Nyboe Tabor Course Outline: Non-Stationary Time Series and Unit Root Testing 1 Stationarity and Deviation from Stationarity Trend-Stationarity

More information

TIME SERIES AND FORECASTING. Luca Gambetti UAB, Barcelona GSE Master in Macroeconomic Policy and Financial Markets

TIME SERIES AND FORECASTING. Luca Gambetti UAB, Barcelona GSE Master in Macroeconomic Policy and Financial Markets TIME SERIES AND FORECASTING Luca Gambetti UAB, Barcelona GSE 2014-2015 Master in Macroeconomic Policy and Financial Markets 1 Contacts Prof.: Luca Gambetti Office: B3-1130 Edifici B Office hours: email:

More information

Ch. 14 Stationary ARMA Process

Ch. 14 Stationary ARMA Process Ch. 14 Stationary ARMA Process A general linear stochastic model is described that suppose a time series to be generated by a linear aggregation of random shock. For practical representation it is desirable

More information

Lecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications

Lecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications Lecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications Prof. Massimo Guidolin 20192 Financial Econometrics Winter/Spring 2018 Overview Moving average processes Autoregressive

More information

Ch. 15 Forecasting. 1.1 Forecasts Based on Conditional Expectations

Ch. 15 Forecasting. 1.1 Forecasts Based on Conditional Expectations Ch 15 Forecasting Having considered in Chapter 14 some of the properties of ARMA models, we now show how they may be used to forecast future values of an observed time series For the present we proceed

More information