Ch 6. Model Specification. Time Series Analysis

Similar documents
Chapter 6: Model Specification for Time Series

Ch 5. Models for Nonstationary Time Series. Time Series Analysis

University of Oxford. Statistical Methods Autocorrelation. Identification and Estimation

Ch 4. Models For Stationary Time Series. Time Series Analysis

STAT Financial Time Series

FE570 Financial Markets and Trading. Stevens Institute of Technology

Part 1. Multiple Choice (50 questions, 1 point each) Part 2. Problems/Short Answer (10 questions, 5 points each)

Univariate ARIMA Models

Stat 5100 Handout #12.e Notes: ARIMA Models (Unit 7) Key here: after stationary, identify dependence structure (and use for forecasting)

Ch 8. MODEL DIAGNOSTICS. Time Series Analysis

Chapter 5: Models for Nonstationary Time Series

Univariate Time Series Analysis; ARIMA Models

Prof. Dr. Roland Füss Lecture Series in Applied Econometrics Summer Term Introduction to Time Series Analysis

Advanced Econometrics

Chapter 8: Model Diagnostics

CHAPTER 8 MODEL DIAGNOSTICS. 8.1 Residual Analysis

Circle a single answer for each multiple choice question. Your choice should be made clearly.

We will only present the general ideas on how to obtain. follow closely the AR(1) and AR(2) cases presented before.

Chapter 4: Models for Stationary Time Series

Ch 9. FORECASTING. Time Series Analysis

Forecasting using R. Rob J Hyndman. 2.4 Non-seasonal ARIMA models. Forecasting using R 1

Lecture 7: Model Building Bus 41910, Time Series Analysis, Mr. R. Tsay

at least 50 and preferably 100 observations should be available to build a proper model

Topic 4 Unit Roots. Gerald P. Dwyer. February Clemson University

Empirical Market Microstructure Analysis (EMMA)

TIME SERIES ANALYSIS AND FORECASTING USING THE STATISTICAL MODEL ARIMA

Estimation and application of best ARIMA model for forecasting the uranium price.

Part 1. Multiple Choice (40 questions, 1 point each) Part 2. Problems/Short Answer (10 questions, 6 points each)

Chapter 12: An introduction to Time Series Analysis. Chapter 12: An introduction to Time Series Analysis

Classic Time Series Analysis

Module 3. Descriptive Time Series Statistics and Introduction to Time Series Models

{ } Stochastic processes. Models for time series. Specification of a process. Specification of a process. , X t3. ,...X tn }

Review Session: Econometrics - CLEFIN (20192)

5 Autoregressive-Moving-Average Modeling

Univariate, Nonstationary Processes

Lesson 13: Box-Jenkins Modeling Strategy for building ARMA models

Autoregressive Moving Average (ARMA) Models and their Practical Applications

The Identification of ARIMA Models

Chapter 9: Forecasting

EASTERN MEDITERRANEAN UNIVERSITY ECON 604, FALL 2007 DEPARTMENT OF ECONOMICS MEHMET BALCILAR ARIMA MODELS: IDENTIFICATION

Lecture 6a: Unit Root and ARIMA Models

APPLIED ECONOMETRIC TIME SERIES 4TH EDITION

A Data-Driven Model for Software Reliability Prediction

STAT 443 Final Exam Review. 1 Basic Definitions. 2 Statistical Tests. L A TEXer: W. Kong

Basics: Definitions and Notation. Stationarity. A More Formal Definition

Chapter 2: Unit Roots

Dynamic Time Series Regression: A Panacea for Spurious Correlations

Problem Set 2: Box-Jenkins methodology

FORECASTING SUGARCANE PRODUCTION IN INDIA WITH ARIMA MODEL

Non-Stationary Time Series and Unit Root Testing

STAT 436 / Lecture 16: Key

Some Time-Series Models

ECON/FIN 250: Forecasting in Finance and Economics: Section 7: Unit Roots & Dickey-Fuller Tests

Autoregressive Integrated Moving Average Model to Predict Graduate Unemployment in Indonesia

Time Series I Time Domain Methods

Econometrics II Heij et al. Chapter 7.1

2. An Introduction to Moving Average Models and ARMA Models

ECONOMETRIA II. CURSO 2009/2010 LAB # 3

Quantitative Finance I

Marcel Dettling. Applied Time Series Analysis SS 2013 Week 05. ETH Zürich, March 18, Institute for Data Analysis and Process Design

6 NONSEASONAL BOX-JENKINS MODELS

Chapter 3: Regression Methods for Trends

Circle the single best answer for each multiple choice question. Your choice should be made clearly.

Lecture 4a: ARMA Model

Time Series 4. Robert Almgren. Oct. 5, 2009

Lecture 2: Univariate Time Series

Applied time-series analysis

AR(p) + I(d) + MA(q) = ARIMA(p, d, q)

Lecture 1: Stationary Time Series Analysis

A SARIMAX coupled modelling applied to individual load curves intraday forecasting

Lecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications

3 Theory of stationary random processes

Financial Time Series Analysis: Part II

Multivariate Time Series Analysis and Its Applications [Tsay (2005), chapter 8]

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley

Forecasting. Simon Shaw 2005/06 Semester II

Time Series 2. Robert Almgren. Sept. 21, 2009

ECON/FIN 250: Forecasting in Finance and Economics: Section 6: Standard Univariate Models

STAT 520: Forecasting and Time Series. David B. Hitchcock University of South Carolina Department of Statistics

Decision 411: Class 9. HW#3 issues

A time series is called strictly stationary if the joint distribution of every collection (Y t

Elements of Multivariate Time Series Analysis

MODELING INFLATION RATES IN NIGERIA: BOX-JENKINS APPROACH. I. U. Moffat and A. E. David Department of Mathematics & Statistics, University of Uyo, Uyo

Univariate linear models

1 Linear Difference Equations

Introduction to ARMA and GARCH processes

Econometrics for Policy Analysis A Train The Trainer Workshop Oct 22-28, 2016 Organized by African Heritage Institution

The ARIMA Procedure: The ARIMA Procedure

Econ 623 Econometrics II Topic 2: Stationary Time Series

Econometrics of financial markets, -solutions to seminar 1. Problem 1

Unit root problem, solution of difference equations Simple deterministic model, question of unit root

AR, MA and ARMA models

Scenario 5: Internet Usage Solution. θ j

Covariances of ARMA Processes

BCT Lecture 3. Lukas Vacha.

Non-Stationary Time Series and Unit Root Testing

Stochastic Modelling Solutions to Exercises on Time Series

Econometría 2: Análisis de series de Tiempo

11. Further Issues in Using OLS with TS Data

SOME BASICS OF TIME-SERIES ANALYSIS

Transcription:

We start to build ARIMA(p,d,q) models. The subjects include: 1 how to determine p, d, q for a given series (Chapter 6); 2 how to estimate the parameters (φ s and θ s) of a specific ARIMA(p,d,q) model (Chapter 7); 3 how to evaluate the fitted model and improve it if needed (Chapter 8). A standard model building process go through the above three procedures in order. We may iterate the process several times to improve the model.

6.1. Properties of the Sample ACF Recall the sample autocorrelation function (sample ACF) for an observed series {Y t } n t=1 is r k = n t=k+1 (Y t Y )(Y t k Y ) n t=1 (Y t Y ) 2, k = 1, 2, 3, We use the sample ACF {r k } to estimate the ACF {ρ k }, then use the pattern of ACF to decide the model. There are three basic questions about the sample ACF: 1 Is sample ACF a unbiased estimator of ACF (E(r k ) = ρ k )? 2 How precise is our estimation (Var (r k ))? 3 What is the correlation Corr (r k, r j ) of r k and r j?

Theorem 1 Suppose {Y t } is a stationary ARMA process. For any fixed m, the joint distribution of n(r1 ρ 1 ), n(r 2 ρ 2 ),, n(r m ρ m ) approaches, as n, a joint normal distribution with zero means and certain covariance matrix [ c ij (see (6.1.2) on pp110.) ]m m In particular, for large n, 1 each r k N(ρ k, c kk n ) (so r k is an unbiased estimator of ρ k with the variance inversely proportional to n); 2 Corr (r k, r j ) c kj ckk c jj is approximately constant for large n.

The sample ACF behaves much simpler for some special cases: If {Y t } is white noise, then Var (r k ) 1 n and Corr (r k, r j ) 0 for k j. If {Y t } AR(1), we get Var (r 1 ) 1 φ2. n The closer φ is to ±1, the more precise our estimate of ρ 1 (= φ) becomes. Var (r k ) 1 [ ] 1 + φ 2 n 1 φ 2 for large k. The closer φ is to ±1, the more variance our estimate of large ρ k φ k is. 1 φ Corr (r 1, r 2 ) 2φ 2 1 + 2φ 2 3φ 4.

Exhibit 6.1 on pp111 gives approximate standard deviations and correlations for several lags and a few values of φ in AR(1) models.

For {Y t } MA(1), we have Var (r k ) c kk /n and Corr (r k, r j ) where c kj ckk c jj c 11 = 1 3ρ 2 1+4ρ 4 1, c kk = 1+2ρ 2 1 for k > 1, c 12 = 2ρ 1 (1 ρ 2 1). Exhibit 6.2 lists large-sample standard deviations and correlations for the sample ACF for several lags and θ-values. The sample ACF can be highly correlated and that the standard deviation of r k is larger for k > 1 than for k = 1.

For {Y t } MA(q), we get Er k = ρ k = 0 for k > q, Var (r k ) = 1 q 1 + 2 ρ 2 j for k > q. (1) n j=1 Theorem 2 A test of the hypothesis that an observed series is MA(q) could be done by comparing r k (k > q) to ±2 standard errors, where the standard error is estimated by Var (r k ) using the above formula and replacing ρ s by r s. We would reject the null hypothesis iff r k lies outside these bounds.

The dashed horizontal lines plotted at ±2/ n are used to test whether or not the ACFs are significantly different from 0. The limits are based on the approximate large sample standard error that applies to a white noise process, namely 1/ n. The sample ACF values exceed these rough critical values at lags 1, 5, and 14.

The same sample ACF with critical bounds calculated by (1). We replace ρ s by r s, let q equal 1, 2, 3,... successively, and take the square root. The sample ACF at lag 5 is barely significant. This plot suggests us to consider a MA(1) model.

As the second example, Exhibit 6.7 displays the sample ACF for an MA(1) Process ma1.2.s with θ = 0.9. The critical values based on the approximately white noise limits point to a MA(1) model. However, the lag 6 sample ACF looks really significant.

If we use the new limits for the sample ACF of the dataset ma1.2.s, the lag 6 sample ACF barely pass the line.

For dataset ma2.s, the sample ACF displays significance at lags 1, 2, 5, 6, 7, 14 and 15 based on the simple standard error bounds.

With the more sophisticated standard error bounds, the lag 7,14, 15 sample ACF of ma2.s is no longer significant. Later we will see that MA(2) is the most appropriate model for these data.

Ex. [HW6.12] From a time series of 100 observations, we calculate r 1 = 0.49, r 2 = 0.31, r 3 = 0.21, r 4 = 0.11; r k < 0.09 for k > 4. On this basis alone, what ARIMA model would we tentatively specify for the series?

6.2. The Partial and Extended Autocorrelation Functions For MA(q) models, ρ k = 0 for k > q. So the ACF/sample ACF is a good indicator for the order of the process. However, the ACF of AR(p) models does not cut off. We need new approaches./new functions to determine the type and the order of AR(p) models. Ex. Consider the AR(1) model Y t = φy t 1 + e t, in which γ 2 = φ 2 γ 0 0 because Y t is dependent on Y t 2 through Y t 1. Suppose we break this chain of dependence by removing (or partial out) the effect of Y t 1. The least square regression of Y t on Y t 1 is ρ 1 Y t 1 = φy t 1. We have Cov (Y t φy t 1, Y t 2 φy t 1 ) = Cov (e t, Y t 2 φy t 1 ) = 0.

6.2. The Partial and Extended Autocorrelation Functions For MA(q) models, ρ k = 0 for k > q. So the ACF/sample ACF is a good indicator for the order of the process. However, the ACF of AR(p) models does not cut off. We need new approaches./new functions to determine the type and the order of AR(p) models. Ex. Consider the AR(1) model Y t = φy t 1 + e t, in which γ 2 = φ 2 γ 0 0 because Y t is dependent on Y t 2 through Y t 1. Suppose we break this chain of dependence by removing (or partial out) the effect of Y t 1. The least square regression of Y t on Y t 1 is ρ 1 Y t 1 = φy t 1. We have Cov (Y t φy t 1, Y t 2 φy t 1 ) = Cov (e t, Y t 2 φy t 1 ) = 0.

6.2.1. PACF Def. The partial autocorrelation at lag k is the correlation between Y t and Y t k after removing the effect of the intervening variables Y t 1, Y t 2,, Y t k+1. 1 If {Y t t } is normally distributed, we can let φ kk = Corr (Y t, Y t k Y t 1, Y t 2,, Y t k+1 ). 2 If we predict Y t and Y t k based on a least square estimation of a linear function of Y t 1,, Y t k+1, we get the predictors Ŷ t = β 1 Y t 1 + + β k 1 Y t k+1, Ŷ t k = β 1 Y t k+1 + + β k 1 Y t 1, (where the coefficients are coincide due to stationarity.) The partial autocorrelation function (PACF) at lag k is then defined to be the correlation between the prediction errors: φ kk = Corr(Y t β 1 Y t 1 β k 1 Y t k+1, Y t k β 1 Y t k+1 β k 1 Y t 1 ).

We now study the PACF for AR(p) and MA(q) models: (AR(p):) For k > p, the best linear predictor of Y t w.r.t. Y t 1, Y t 2,, Y p,, Y t k+1 is φ 1 Y t 1 + φ 2 Y t 2 + + φ p Y t p (Chapter 9). Let Ŷt p be the linear predictor of Y t p w.r.t. Y t 1, Y t 2,, Y t k+1. Then ( ) φ kk = Corr Y t φ 1 Y t 1 φ p Y t p, Y t p Ŷt p ( ) = Corr e t, Y t p Ŷt p = 0. Theorem 3 The PACF of an AR(p) model satisfies that φ kk = 0 for k > p.

(MA(1) & MA(q):) We will get φ kk = θk (1 θ 2 ), for k 1. 1 θ2(k+1) The PACF of an MA(1) model decays to zero exponentially. In general, the PACF of an MA(q) model behaves like the ACF of an AR(q) model.

Theorem 4 For any stationary process with given ACF {ρ i }, the PACF {φ kk } could be solved by Yule-Walker equations: ρ j 1 φ k1 + ρ j 2 φ k2 + ρ j 3 φ k3 + + ρ j k φ kk = ρ j for j = 1, 2,, k. (2) Explicitly, φ k1 + ρ 1 φ k2 + ρ 2 φ k3 + + ρ k 1 φ kk = ρ 1 ρ 1 φ k1 + φ k2 + ρ 1 φ k3 + + ρ k 2 φ kk = ρ 2. ρ k 1 φ k1 + ρ k 2 φ k2 + ρ k 3 φ k3 + + φ kk = ρ k We solve φ k1, φ k2,, φ kk and discard all but φ kk. For AR(p) model and k = p, the parameters φ 1,, φ p also satisfies the Yule-Walker equations. We have φ pp = φ p 0 and φ kk = 0 for k > p. Thus PACF effectively display the correct order of an AR(p) process.

6.2.2 Sample PACF Definition 5 In the Yule-Walker equations (2), we estimate the ACF ρ s by the sample ACF r s, and solve the resulting linear equations for k = 1, 2, 3,... to get estimates of φ kk. The estimated function is called the sample partial autocorrelation function (sample PACF), denoted by ˆφ kk. The Yule-Walker equations (2) can be solved recursively: For an observed series, the sample PACF { ˆφ kk } can be estimated by replacing ρ s by r s in the above formula.

Under the hypothesis that an AR(p) model is correct, for k > p, ˆφ kk is approximately N(0, 1/n). Thus for k > p, ±2/ n can be used as critical limits on ˆφ kk. The ACF shows strong positive correlations at lags 1, 2, 3, as seen early. However, the sample ACF decreases more linearly than exponentially as theory suggests. Also contrary to theory, the sample ACF goes negative at lag 10 and remains so for many lags.

The sample PACF gives a much clearer picture and strongly suggests an AR(1) model.

The sample ACF looks like the damped wave. However, it does not damp down nearly as quickly as theory predicts.

The sample PACF gives a strong indication that we should consider an AR(2) model for these data. The sample PACF at lag 9 exceeds critical line and it would need to be investigated further during model diagnostics.

6.2.3 Mixed Models and the Extended ACF The behavior of ACF and PACF are summarized in Exhibit 6.3. The sample ACF and PACF provide effective tools for identifying pure AR(p) or MA(q) models.

For a mixed ARMA model, many graphical tools have been proposed to effectively identify the ARMA orders: the corner method (Becuin et al., 1980), the extended autocorrelation (EACF) method (Tsay and Tiao, 1984), and the smallest canonical correlation (SCAN) method (Tsay and Tiao, 1985), among others. The EACF method seems to have good sampling properties for moderately large sample sizes (W. S. Chan, 1999). The EACF method uses the fact that if the AR part of a mixed ARMA model is known, filtering out the autoregression from the observed time series results in a pure MA process that enjoys the cutoff property in its ACF. The AR coefficients may be estimated by a finite sequence of regressions.

Ex. Consider the ARMA(1,1) model: Y t = φy t 1 + e t θe t 1. Recall ρ 1 = (φ θ)(1 φθ) 1 2φθ + θ 2, ρ 2 = φρ 1. 1 Do a linear regression of Y t on Y t 1 gets ρ 1 Y t 1. The residual R (1) t is not a MA(1) process. := Y t ρ 1 Y t 1 = (φ ρ 1 )Y t 1 + e t θe t 1 2 Do a second multiple regression of Y t on Y t 1 and on R (1) t 1 = Y t 1 ρ 1 Y t 2, the lag 1 of the residuals from the first regression: Y t = φy t 1 + µ(y t 1 ρ 1 Y t 2 ). By least squares property, the correlations of Y t φy t 1 µ(y t 1 ρ 1 Y t 2 ) with Y t 1 and Y t 1 ρ 1 Y t 2 (equivalently, with Y t 2 ) should be zero. We solve that φ = φ, µ = ρ 1 φ 1 ρ 2. 1 The coefficient φ of Y t 1 in the second regression is a consistent estimator of φ. 3 The process W t = Y t φy t 1 is approximately an MA(1) process.

Ex. Consider the ARMA(1,1) model: Y t = φy t 1 + e t θe t 1. Recall ρ 1 = (φ θ)(1 φθ) 1 2φθ + θ 2, ρ 2 = φρ 1. 1 Do a linear regression of Y t on Y t 1 gets ρ 1 Y t 1. The residual R (1) t is not a MA(1) process. := Y t ρ 1 Y t 1 = (φ ρ 1 )Y t 1 + e t θe t 1 2 Do a second multiple regression of Y t on Y t 1 and on R (1) t 1 = Y t 1 ρ 1 Y t 2, the lag 1 of the residuals from the first regression: Y t = φy t 1 + µ(y t 1 ρ 1 Y t 2 ). By least squares property, the correlations of Y t φy t 1 µ(y t 1 ρ 1 Y t 2 ) with Y t 1 and Y t 1 ρ 1 Y t 2 (equivalently, with Y t 2 ) should be zero. We solve that φ = φ, µ = ρ 1 φ 1 ρ 2. 1 The coefficient φ of Y t 1 in the second regression is a consistent estimator of φ. 3 The process W t = Y t φy t 1 is approximately an MA(1) process.

Ex. Consider the ARMA(1,1) model: Y t = φy t 1 + e t θe t 1. Recall ρ 1 = (φ θ)(1 φθ) 1 2φθ + θ 2, ρ 2 = φρ 1. 1 Do a linear regression of Y t on Y t 1 gets ρ 1 Y t 1. The residual R (1) t is not a MA(1) process. := Y t ρ 1 Y t 1 = (φ ρ 1 )Y t 1 + e t θe t 1 2 Do a second multiple regression of Y t on Y t 1 and on R (1) t 1 = Y t 1 ρ 1 Y t 2, the lag 1 of the residuals from the first regression: Y t = φy t 1 + µ(y t 1 ρ 1 Y t 2 ). By least squares property, the correlations of Y t φy t 1 µ(y t 1 ρ 1 Y t 2 ) with Y t 1 and Y t 1 ρ 1 Y t 2 (equivalently, with Y t 2 ) should be zero. We solve that φ = φ, µ = ρ 1 φ 1 ρ 2. 1 The coefficient φ of Y t 1 in the second regression is a consistent estimator of φ. 3 The process W t = Y t φy t 1 is approximately an MA(1) process.

Similarly, the AR coefficients of an ARMA(p,q) model can be consistently estimated via a sequence of q regressions. The AR and MA orders p and q can be determined by an iterative procedure. Let W t,k,j = Y t φ 1 Y t 1 φ k Y t k be the autoregressive residuals defined with the AR coefficients estimated iteratively assuming the AR order is k and the MA order is j. The sample autocorrelations of W t,k,j for t are called the sample EACFs. For k = p and j q, {W t,k,j } t MA(q), so that its autocorrelations of lag q + 1 or higher are equal to zero. For k > p, an overfitting problem occurs, and this increases the MA order for the W process by the minimum of k p and j q.

We can summarize the information in a table to determine p and q.

We simulate an series of mixed ARMA(1,1) model and analyze the data.

The sample ACF and the sample PACF seem to indicate that an AR(1) model should be specified. However, the triangular region of zeros shown in the sample EACF in Exhibit 6.17 indicates quite clearly that a mixed model with q = 1 and with p = 1or2 would be more appropriate.

You may simulate and analyze a ARMA(1,1) series by yourself. (See TS-ch6.R)

6.3 Specification of Some Simulated Time Series The exhibits have been illustrated in 6.1 and 6.2. 6.4 Nonstationarity The definition of sample ACF implicitly assumes stationarity. However, for nonstationary series, the sample ACF often fails to die out rapidly as the lags increase, due to the tendency for nonstationary series to drift slowly.

6.3 Specification of Some Simulated Time Series The exhibits have been illustrated in 6.1 and 6.2. 6.4 Nonstationarity The definition of sample ACF implicitly assumes stationarity. However, for nonstationary series, the sample ACF often fails to die out rapidly as the lags increase, due to the tendency for nonstationary series to drift slowly.

The sample ACF values are significantly far from zero and they decrease slowly. The sample ACF is unlike that of a stationary one. Similar pattern happens for the series log(oil.price). The PACF of oil.price seems strongly inclined to an AR(1) model. However, the plot of oil.price has shown that it is nonstationary.

The plot of diff(log(oil.price)) supports stationarity with possible outliners. The ACF looks like a damping sine wave with significant lag 1 value. Therefore, an MA(1) model for diff(log(oil.price)) seems appropriate. The EACF also supports an MA(1) model. Thus we may assign diff (log(oil.price)) MA(1), log(oil.price) IMA(1, 1).

In the difference approach, we difference the observed series until the plot of resulting series and its sample ACF, PACF, or EACF supports a stationary process. Usually we will achieve stationarity after one or two differences, together with logarithm or some other transforms. 6.4.1 Overdifferencing The difference of a stationary series is also stationary. However, overdifferencing introduces unnecessary correlations into a series.

In the difference approach, we difference the observed series until the plot of resulting series and its sample ACF, PACF, or EACF supports a stationary process. Usually we will achieve stationarity after one or two differences, together with logarithm or some other transforms. 6.4.1 Overdifferencing The difference of a stationary series is also stationary. However, overdifferencing introduces unnecessary correlations into a series.

Ex. Suppose an observed series {Y t } ARIMA(p, d, q). Then d Y t ARMA(p, q). If we overdifference the series and get d+1 Y t ARMA(p, q + 1) then we will mistakenly model the original series as {Y t } ARIMA(p, d + 1, q + 1). Drawbacks: 1 Overdifferecing makes the model more complicated. 2 Overdifferecing creates a noninvertible model, which will create serious problems when we attempt to estimate their parameters.

If we overdifference the dataset rwalk, then we may model 2 (rwalk) MA(1) and mistakenly model rwalk IMA(2, 1).

The right difference suggests that (rwalk) e t and thus rwalk ARIMA(0, 1, 0) (a random walk process).

Remember the principle of parsimony: models should be simple, but not too simple.

6.4.2 The Dickey-Fuller Unit-Root Test (not required) Consider the model Y t = βy t 1 + X t, t = 1, 2, where X t is stationary. Let a = 1 β. The series {Y t } is nonstationary iff β = 1 or a = 0. We set up the hypothesis testing: Null hypothesis: a = 0 Alternative hypothesis: a < 0. The test for differencing amounts to testing for a unit root in the AR characteristic polynomial of {Y t }. The augmented Dickey-Fuller (ADF) test statistic is the t-statistic of the estimated coefficient of a from the method of least squares regression. However, the ADF test statistic is not approximately t-distributed under the null hypothesis; instead, it has a certain nonstandard large-sample distribution under the null hypothesis of a unit root. We skip the details here.

6.5 Other Specification Methods Some popular approaches to determine the ARMA orders p and q: 1 Akaike s Information Criterion (AIC): to select the model that minimizes AIC = 2 log(maximal likelihood) + 2k where k = p + q + 1 if the model contains an intercept or constant term and k = p + q otherwise. The AIC is an estimator of the average Kullback-Leibler divergence of the estimated model from the true model. 2 The Schwarz Bayesian Information Criterion (BIC): to select the model that minimizes BIC = 2 log(maximal likelihood) + k log(n).

A subset ARMA(p,q) model is an ARMA(p,q) model with a subset of its coefficients known to be zero. For a subset ARMA(12,12) model like Y t = 0.8Y t 12 + e t + 0.7e t 12, finding a subset ARMA model that adequately approximates the underlying process is more important than simply determining the ARMA orders. It is prudent to examine a few best subset ARMA models (in terms of, for example, BIC) in order to arrive at some helpful tentative models for further study. We show an example of the above model in the next exhibit.

Exhibit 6.22 Best Subset ARMA Selection Based on BIC Each row in the exhibit corresponds to a subset ARMA model where the cells of the variables selected for the model are shaded. The models are sorted according to their BIC, with better models (lower BIC) placed in higher rows and with darker shades. The best Subset ARMA selection only suggests some possible models. These models need further examinations. In the above exhibit, if we use set.seed(3399) or some other numbers, we may have different outcomes.

Exhibit 6.22 Best Subset ARMA Selection Based on BIC Each row in the exhibit corresponds to a subset ARMA model where the cells of the variables selected for the model are shaded. The models are sorted according to their BIC, with better models (lower BIC) placed in higher rows and with darker shades. The best Subset ARMA selection only suggests some possible models. These models need further examinations. In the above exhibit, if we use set.seed(3399) or some other numbers, we may have different outcomes.

6.6 Specification of Some Actual Time Series The Los Angeles Annual Rainfall Series larain Taking logarithms significantly improves the normality. You may also try shapiro.test to check.

6.6 Specification of Some Actual Time Series The Los Angeles Annual Rainfall Series larain Taking logarithms significantly improves the normality. You may also try shapiro.test to check.

Based on the sample ACF, we could model the logarithm of annual rainfall amount as independent, normal random variables with mean 2.58 and standard deviation 0.478.

The Chemical Process Color Property Series color The ACF plot looks like an MA(1) model. However, there is a damp sine wave pattern. Maybe an AR or ARMA model is more appropriate.

The PACF plot suggests an AR(1) model.

The Annual Abundance of Canadian Hare Series hare The confidence interval contains λ = 0.5. We will take square roots on hare for further analysis.

The damped sine wave pattern motivate us to look for PACF.

The PACF plot suggests an AR(2) or AR(3) model.

The Oil Price Series oil.price This table suggests an ARMA model with p = 0 and q = 1, that is, MA(1).

For Y t := log(oil t ), the best model is in terms of Y t 1 and Y t 4 and no lags are needed in error terms. The second best model is AR(1) (where log(oil t ) ARI (1, 1)).

The plot suggests an MA(1) model for Y t := log(oil t ). The sample ACF seems to have some damped sine wave pattern.

The plot suggests an AR(2) model for Y t := log(oil t ). We will investigate all the above models further.

The plot suggests an AR(2) model for Y t := log(oil t ). We will investigate all the above models further.