Improved model selection criteria for SETAR time series models

Size: px
Start display at page:

Download "Improved model selection criteria for SETAR time series models"

Transcription

1 Journal of Statistical Planning and Inference Improved model selection criteria for SEAR time series models Pedro Galeano a,, Daniel Peña b a Departamento de Estadística e Investigación Operativa, Universidad de Santiago de Compostela, 5782 Santiago de Compostela, Spain b Departamento de Estadística, Universidad Carlos III de Madrid, Getafe, Madrid, Spain Received 20 February 2006; received in revised form 8 September 2006; accepted 6 October 2006 Available online February 2007 Abstract he purpose of this paper is threefold. First, we obtain the asymptotic properties of the modified model selection criteria proposed by Hurvich et al Improved estimators of Kullback-Leibler information for autoregressive model selection in small samples. Biometrika 77, for autoregressive models. Second, we provide some highlights on the better performance of this modified criteria. hird, we extend the modification introduced by these authors to model selection criteria commonly used in the class of self-exciting threshold autoregressive SEAR time series models. We show the improvements of the modified criteria in their finite sample performance. In particular, for small and medium sample size the frequency of selecting the true model improves for the consistent criteria and the root mean square error RMSE of prediction improves for the efficient criteria. hese results are illustrated via simulation with SEAR models in which we assume that the threshold and the parameters are unknown Elsevier B.V. All rights reserved. Keywords: Asymptotic efficiency; Autoregressive models; Consistency; Model selection criteria; SEAR models. Introduction Since the seminal work by Akaike 969, model selection criteria have become a widely used tool for selecting the order of different time series models. Most of these criteria can be classified into two groups. he first one includes the efficient criteria, which asymptotically select the model which produces the least mean square prediction error. he final prediction error criterion FPE, by Akaike 969, the Akaike information criterion AIC, by Akaike 973 and the corrected Akaike information criterion AICc, by Hurvich and sai 989 are efficient criteria. he FPE selects the model that minimizes the one step ahead square prediction error. he AIC is an estimator of the expected Kullback Leibler divergence between the true and the fitted model, while the AICc is a bias correction form of the AIC that appears to work better in small samples. he second group includes the consistent criteria, which, under the assumption that the data come from a finite order autoregressive process, asymptotically select the true order of the process. he Bayesian information criterion BIC, by Schwarz 978, and the Hannan Quinn criterion HQC, by Hannan and Quinn 979, are consistent criteria. he BIC approaches the posterior probabilities of the models, while the HQC is designed to be a consistent criterion with the fastest convergence rate to the true model. Corresponding author. el.: x3207; fax: address: pgaleano@usc.es P. Galeano /$ - see front matter 2007 Elsevier B.V. All rights reserved. doi:0.06/j.jspi

2 P. Galeano, D. Peña / Journal of Statistical Planning and Inference All these criteria can be written compactly as members of the family of criteria: min { p log σ2 p + p + C,p + }, where p is the order of the autoregressive process, σ 2 p is the maximum likelihood estimate of the residual variance of the process, is the sample size and C,p + = p+ +p+ log p+ for the FPE, C,p + = 2 for the AIC, C,p + = for the AICc, C,p + = log, for the BIC and C,p + = 2m log log with 2 p+ m>, for the HQC. Hurvich et al. 990 further approximated the expected Kullback Leibler divergence to derive a criterion, AICc, which can be written as follows: min p { log Σ α p + 2p+ p + }, 2 where Σ α p is the determinant of the estimated covariance matrix of the series under the autoregressive process with order p and parameters α p, that will be defined in Section 2. hese authors also introduced the AIC and BIC criteria, by replacing log σ 2 p by the determinant term in the AIC and BIC criteria, and showed in a Monte Carlo experiment the good performance on this modification. However, they did not study the asymptotic properties of these modified criteria. he first contribution of this paper is to show that the asymptotic properties of the original criteria applies to the modified criteria 2. hus, we show the efficiency of AIC and AICc and the consistency of BIC. Although Hurvich et al. 990 showed via simulation the better performance of the modified criteria, no theoretical reasons have been given explaining this improvements. he second contribution of this paper is to provide three interpretations on the advantages of using the determinant term by using three different comparisons: the one step ahead prediction variances; 2 the correlation structure; 3 a measure of the goodness of the fit. A useful nonlinear extension of linear time series models are the self-exciting threshold autoregressive SEAR models, see ong 990. hese models can explain interesting features found in real data, such as asymmetric limit cycles, jump phenomena, chaos and so on. Model selection for SEAR models has been addressed in several papers. ong 990 suggested to use the AIC but no theoretical justification was given. Wong and Li 998 showed that the AICc criterion is an asymptotically unbiased estimator of the expected Kullback Leibler information for SEAR models and analyzed the small sample properties of AIC, AICc and BIC via simulation experiments. Kapetanios 200 extended some of the existing theoretical results for several model selection criteria in linear models to threshold models. De Gooijer 200 proposed three cross-validation criteria. Campbell 2004 and Unnikrishnan 2004 developed Bayesian model selection within a Markov Chain Monte Carlo MCMC framework, and, finally, Öhrvik and Schoier 2005 studied the performance of several bootstrap selection criteria. As a SEAR model is piecewise autoregressive linear, it seems natural to extend the modification considered by Hurvich et al. 990 for autoregressive models to these nonlinear models. he third contribution of this paper is to present new SEAR model selection criteria based on the determinant term and show via a Monte Carlo study the better performance for small and medium sample size of these modified criteria. he rest of this paper is organized as follows. Section 2 briefly reviews model selection criteria for the class of linear autoregressive models, proves that the correction by the determinant term keeps their asymptotic properties, and provides some intuition to justify why this correction can improve their performance. Section 3 develops the modification by the determinant term for SEAR time series model selection criteria. Section 4 shows the better performance of the modified criteria in a Monte Carlo experiment. 2. Model selection criteria for the class of linear autoregressive processes 2.. Model selection criteria for autoregressive processes Suppose it is known that a given time series, x = x,...,x, has been generated by the class of autoregressive AR Gaussian processes, given by x t x t p x t p = a t, t =...,, 0,,...,

3 2804 P. Galeano, D. Peña / Journal of Statistical Planning and Inference where a t is a sequence of independent Gaussian distributed random variables with zero mean and variance σ 2 p.we assume that p {0,...,p max }, where p max is an upper bound. he ARp model, denoted by M p, has parameters α p = β p, σ2 p, where β p is the p max vector of parameters, β p = p,..., pp, 0,...,0. }{{} p max We assume that the models M p are causal, invertible and stationary. hen the covariance matrix of x can be written as Σα p = σ 2 p Qβ p, where Qβ p is a matrix which only depends on β p. Let, Qβ p = Lβ p L β p be the Cholesky decomposition of Qβ p such that aβ p = Lβ p x. We denote the parameters of the model that have generated the data as α 0 = β 0, σ2 0. In practice, the model parameters are unknown. he maximum likelihood estimates of the vector of parameters β p and the innovations variance σ 2 p, denoted by β p and σ 2 p, respectively, are obtained after maximizing the likelihood function, px M p, given by: px M p = 2π /2 Σα p /2 exp 2 x Σ α p x. 3 Akaike 973 proposed to select the model which minimizes the expected Kullback Leibler divergence between the fitted and the true model, defined by, [ [ E αp E α0 2 log py α ]] 0 = E αp [E α0 [ 2 log py α p ]] E αp [E α0 [ 2 log py α 0 ]] 4 py α p for an arbitrary realization y = y,...,y of the process. As 4 is always positive, minimizing it implies making py α p as close as possible to py α 0, in the expected Kullback Leibler divergence. As the second term of the expected Kullback Leibler divergence is constant for all the models, minimizing 4 is equivalent to minimize, [ ] E αp [E α0 [ 2 log py α p ]] = 2 log py α p py α 0 dy p α p α 0 d α p, 5 where y and α p are assumed to be independent. hus, the rule proposed by Akaike selects the autoregressive model that minimizes 5 with respect to the two sources of uncertainty: the distribution of future observations given the parameters and the distribution of the estimate. Akaike 973 approached 5 as follows: E αp [E α0 [ 2 log py α p ]] = log 2π + + log σ 2 p + 2p + + o p, which leads to the AIC AICp = log σ 2 p + 2p +. Hurvich and sai 989 obtained an approximation of 5 which reduces the small sample bias of the approximation by Akaike 973, and is given by E αp [E α0 [ 2 log py α p ]] = log 2π + + log σ 2 p + 2p+ p + + o p which leads to the AICc AICcp = log σ 2 p + 2p+ p +. From the Bayesian point of view, the model selected is the one with maximum posterior probability, pm p x, where pm p x px M p pm p, and px M p = px α p,m p pα p M p dα p, such that pm p and pα p M p are the

4 P. Galeano, D. Peña / Journal of Statistical Planning and Inference prior probabilities of models and parameters, respectively. Schwarz 978 approximated the posterior probability pm p x to derive the BIC, given by BICp = log σ 2 p + log p +. he criteria AIC, AICc and BIC can be written in a compact way as members of the family of criteria, min { p log σ2 p + p + C,p + }, 6 2 where C,p + is 2, for AIC, p+, for AICc, and log, for BIC. Hurvich et al. 990 noted that the expected Kullback Leibler divergence was better approximated if log σ 2 p is replaced by the term log Σ α p and defined the AIC p, AICc p and BIC p criteria, which may be written in compact form as: min p {log Σ α p +p + C,p + }. 7 hese authors showed by simulation the advantages of considering log Σ α p instead of log σ 2 p for autoregression fitting, but did not analyze the asymptotic properties of the modified criteria. First, we show that the AIC and AICc criteria are efficient in the following heorem, whose proof is in Appendix A. heorem. Assume that the following assumptions hold: A {x t } is generated by a stationary process x t x t 2 x t 2 =a t, t =,, 0,, where a t is a sequence of independent Gaussian distributed random variables with zero mean and variance σ 2 a and j= j < ; A2 the polynomial z = z 2 z 2, is nonzero for every complex number z with z ; A3 the upper bound p max is a sequence of positive integers which depends on such that p max and p max / 0 as ; A4 {x t }is not degenerate to a finite order autoregressive process. hen, the AIC and AICc are efficient criteria. On the other hand, the consistency property of the BIC criterion is established in the following heorem, proved in Appendix A. heorem 2. Assume that the following assumptions hold: B {x t } is generated by a stationary process x t x t p x t p = a t, t =,, 0,,...where a t is a sequence of independent Gaussian distributed random variables with zero mean and variance σ 2 a ; B2 the polynomial z = z 2 z 2 p z p, is nonzero for every complex number z with z ; B3 the upper bound p max is fixed and known a priori. hen, the BIC is a consistent criteria. hus, AIC, AICc and BIC are similar to AIC, AICc and BIC for large samples but as we will see in the Monte Carlo study in Section 4, in small and medium sample settings, the difference between the performance of these criteria may be substantial hree interpretations on the advantages of using AIC, AICc and BIC Next, we provide three interpretations on the advantages of using the determinant term Σ α p by showing that this term leads to a better comparison of the models under consideration by the one step ahead prediction variances; 2 the correlation structure; 3 a goodness of fit test Interpretation by one step ahead prediction variances o show the first interpretation, let x t p be the one step ahead mean square predictions under the M p model and let e t p = x t x t p be the corresponding one step ahead prediction errors. hese errors have variances which can be written as E[e t p 2 ]=σ 2 p v2 t p see, Harvey, 98. For instance, for the AR model, v2 t = / 2, for t =,

5 2806 P. Galeano, D. Peña / Journal of Statistical Planning and Inference and vt 2 p =, for t>. hus, the logarithm of the determinant of the covariance matrix of x can be written as / log Σ α p = log σ 2 p v2 t p = log σ 2 p v2 t p, where v 2 t p are obtained after replacing the estimated parameters in v2 t p, and log Σ α p is times the logarithm of the geometric mean of the estimated one step ahead prediction variances. herefore, the difference AIC p + AIC p can be written as AIC p + AIC p = log σ2 p+ v2 t p + σ p v2 t p he first term measures the relative change between the one step ahead prediction variances, while the second term is a penalization for the inclusion of one additional parameter. herefore, AIC p + <AIC p if the geometric mean of the one step ahead prediction variances under the model M p+ is significatively smaller than the corresponding mean under the model M p, or in other words, the AIC will select the model which has a better predictive performance penalized by the number of parameters. As AIC p = log σ 2 p + log v t 2 p + 2p + = AICp + log v t 2 p, the AICp does not take into account the terms v t 2p, but only the estimated residual variance, σ2 p. he same conclusions holds for the AICc and BIC criteria Interpretation by the correlation structure Now we show that the determinant term provides a more sophisticated comparison of the autocorrelation structure of the models under consideration. As Qβ p = σ 2 x /σ2 p Rα p, where σ 2 x is the variance of x and Rα p is the correlation matrix of x, wehave Qβ p =σ 2 x /σ2 p Rα p. Durbin 960 and Ramsey 974 showed that p σ 2 x /σ2 p = p 2 ii β p, Rα p = 2 ii β p i, respectively, where ii β p are the partial autocorrelations of the process under the model M p, so that, Qβ p = p 2 ii β p i p 2 ii β p = p 2 ii β p i. Consequently, the criteria 7 can be written as follows: { p } min log σ 2 p p + p + C,p + i log 2 ii β p, while the difference AIC p + AIC p is AIC p + AIC p = log σ2 p+ σ 2 p + 2 p i log 2 ii β p+ 2 ii β p p + log 2 p+,p+ β p+. he first two terms are the one used by the AIC criterion but now two additional terms appear in the comparison. hey measure the discrepancy between all the partial autocorrelation coefficients under both hypothesis, M p and M p+, with weights that increase with the lag. herefore, AIC p + <AIC p if either: a σ 2 p+ is smaller enough than σ2 p

6 P. Galeano, D. Peña / Journal of Statistical Planning and Inference or b the weighted sum of the partial autocorrelation coefficients computed under the ARp model is greater enough than the corresponding sum under the M p+ model. Note that the last term is acting as a penalization term because it is always positive. he same interpretation applies to BIC and AICc, and the only difference is the penalization term for including one additional parameter Interpretation by a goodness of fit test A last interpretation is given in terms of goodness of fit tests for time series. Peña and Rodriguez 2006 used the log of the determinant of the autocorrelation matrix of the estimated residuals for testing goodness of fit in time series. As aβ p = Lβ p x, we can write [ ] aβ paβ p =[Lβ p ] xx [Lβ p ]. 8 hus, after fitting the model M p, a β p a β p = σ 2 p Ra β p, 9 where Ra β p is the sample correlation matrix of the estimated residuals, a β p. herefore, from 8 and 9, we have so that, log σ 2 p + log Q β p =log / xx log Ra β p AIC p + AIC p = log Ra β p log Ra β p+ +2. hus, AIC p + <AIC p if the logarithm of the determinant of the correlation matrix of the estimated residuals after the model M p+ is significative larger than the one for the model M p. he terms log Ra β p and log Ra β p+ are the values of the statistic proposed by Peña and Rodriguez 2006 for models M p and M p+. herefore, the AIC will select the model with have a significatively larger value of the statistic proposed by Peña and Rodriguez Consequently, the term log σ 2 p + log Q β p, can be seen as a measure of the goodness of fit of the model M p to the series x. As before, the same interpretation applies to BIC and AICc after changing the penalization term. 3. Model selection in SEAR models One of the most often used nonlinear time series model is the SEAR model. A time series data, x = x,...,x, generated by the class of SEAR processes follows the model: p j x t = j0 + ji x t i + a jt, if r j x t d <r j, j =,...,k, 0 where a jt are sequences of independent Gaussian distribution random variables with zero mean and variances σ 2 j.we assume that p j {0,...,pj max } and d {0,...,d max } are nonnegative integers, where pj max, j =,...,kand d max, are some upper bounds, and = r 0 <r < <r k <r k = are the thresholds. he SEARp,...,p k,dmodel, denoted by M p,...,p k,d, has parameters α p,...,p k,d = β p,...,p k,d, σ2,...,σ2 p k, where β p,...,p k,d is the k j= pj max + + k + 2 vector of parameters, β p,...,p k,d = 0,,..., p, 0,...,0, k0, k,..., kpk, 0,...,0,r 0,...,r }{{}}{{} k, d. k j= pj max k+ +

7 2808 P. Galeano, D. Peña / Journal of Statistical Planning and Inference We assume that the models M p,...,p k,d are stationary, ergodic with finite second moments and the stationary distribution of x = x,...,x admits a density that is positive everywhere. We denote the parameters of the model that have generated the data as α 0. Exact maximum likelihood estimates of the parameters of model 0 are not considered because of the complexity of the likelihood function. Assuming that d is known, the conditional log-likelihood of model 0 is given by log p c x α p,...,p k,d = 2 log2π 2 k j log σ 2 p j + S j,r j,r j σ 2, p j j= where j are the number of observations in each regime for the thresholds r,...,r k, j = j0,..., jpj and S j,r j,r j = r j x t d <r j ajt 2, j =,...,k. he conditional maximum likelihood estimators of the parameters, denoted by α p,...,p k,d = β p,...,p k,d, σ2,..., σ2 p k, are the values that maximize the conditional likelihood in, with residual variances σ 2 j = S j, r j, r j j, j =,...,k. Chan 993 showed the strong consistency of the conditional least squares estimators of the parameters for k =. Wong and Li 998 approximated the expected Kullback Leibler divergence for SEAR models as follows: E αp,...,p k,d [E α 0 [ 2 log p c y α p,...,p k,d]] = log2π + k i log σ 2 p i + k i i + p i + i p i 3 + o p, which leads to the AICc criterion for SEAR models. hese authors compared in a simulation study three model selection criteria, AIC, AICc and BIC, which for k regimes are given by min p,...,p k k [ j log σ 2 j + p j + C j j,p j + ], 2 j= j j +p j + where C j j,p j + is 2 for AIC, p j + j p j + 2 for AICc, and log j for BIC. he procedure proposed by Wong and Li 998 works as follows, when k = 2, r = r and d are unknown: a fix the maximum autoregressive and delay orders {p max,p2 max,d max }; b assume r [l,u] R, where l is the % percentile and u is the % percentile of x; c let x,...,x be the order statistics of x; d let I r ={[0.25 ],...,[0.75 ]}, where [ν ] is the largest integer less than ν. Set r = x i, i I r ; e calculate min p,p 2,d,x i {MSCp,p 2,d,x i }, where MSCp,p 2,d,x i is one of the model selection criteria in 2. he autoregressive orders p,p 2, the delay parameter, d, and the threshold, x i, selected are the ones that minimize MSCp,p 2,d,x i. Wong and Li 998 carried out a Monte Carlo experiment for different models and sample sizes for the criteria in 2, and conclude that the AICc has the best performance for small sample sizes and BIC for medium and large sample sizes. De Gooijer 200 proposed a procedure for selecting and estimating the parameters of a SEAR model by crossvalidation which works as follows: a d are as in the previous procedure; e omit one observation of the series, x t, and with the remaining data set obtain conditional least squares estimates of the parameters of the corresponding model, which we denote by t j, predict the omitted observation and obtain the predictive residual, a t t j, r j, r j ; f repeat the previous step for all the observations. he final model is the one that minimizes one of the criteria C,Ccand

8 P. Galeano, D. Peña / Journal of Statistical Planning and Inference Cu, written in compact way as follows: min p,...,p k log at 2 t j, r j, r j + k p j + C j j,p j +, 3 j= where C j j,p j + is 0 for C, j j +p j + p j + j p j 3 for Cc, and p j + [ j j +p j + j p j 3 + j log{ j j p j 2 }] for Cu. he C criterion was analyzed in Stoica et al. 986 for linear models and proved that for a given model, C =AIC+O /2. he Cc and the Cu criteria came from adding the penalty terms of the AICc and AICu criteria to the C criterion. he AICu is a criterion introduced by McQuarrie et al. 997 for linear models and is neither efficient nor consistent but it has a good performance in finite samples. For SEAR models the AICu criterion can be written as in 2 with C j j,p j + = p j + [ j j +p j + j p j 3 + j log{ j j p j 2 }]. As SEAR models are piecewise autoregressive linear, we propose to modify the criteria as in the autoregressive models case as follows. he criteria BIC, AIC, AICc and AICu are modified by adding the determinant term in each regime as follows: k min [ j log σ 2 j + p j + C j j,p j + + log Q j ]. 4 p,...,p k j= In order to compute the determinant term in each regime we first estimate the parameters of the model by conditional likelihood and then obtain the determinant term in each regime. For that we use the expression provided by Leeuw 994 who showed that Q j = M M NN, 5 where M and N are p j p j matrices with elements given by 0 a<b, M ab = a = b, a>b, j,a b { j,pj N ab = +a b, a b, 0, a>b. In the same way, we modify the cross-validation criteria C, Cc and Cu proposed by De Gooijer 200 in 3 by adding the determinant term in each regime. herefore, the modified cross-validation criteria C,Cc and Cu are defined as follows: min p,...,p k log at 2 t j, r j, r j + k [p j + C j j,p j + + log Q j ]. 6 j= he procedures in Wong and Li 998 and De Gooijer 200 are modified by adding the determinant term in the last step obtained with the conditional least squares estimates of the parameters with the whole series. hen, the final model selected is the one that minimizes one of the criteria in 6. he determinant term is computed using Monte Carlo experiments o evaluate the performance of the proposed criteria for SEAR models, 000 realizations were generated from the following two stationary SEAR models: { xt = 0.8x M t + a t, x t 0, x t = 0.2x t + a 2t, x t > 0, M2 { xt = 0.5x t + a t, x t 0, x t = 0.5x t + a 2t, x t > 0,

9 280 P. Galeano, D. Peña / Journal of Statistical Planning and Inference able Frequency of times of correct selection, root mean square errors of the threshold parameter and root mean square prediction errors assuming that d is known M = 30 BIC BIC AIC AIC AICc AICc AICu AICu C C Cc Cc Cu Cu p,p RMSE RMSPE p,p RMSE RMSPE = 50 p,p RMSE RMSPE p,p RMSE RMSPE = 00 p,p RMSE MSPE p,p RMSE RMSPE where a jt N0,, j =, 2. Based on Section 3, we compare the performance of the criteria in 2 with respect to the criteria in 4 and the criteria in 3 with respect to the criteria in 6. In all cases, 000 series were generated from models M and M2 with sample sizes = 3, 5 and 0. We proceed as in Wong and Li 998 and De Gooijer 200 by using a grid to estimate the threshold parameter r. We fit each model to the first observations of each series by conditional likelihood and obtain the determinant term in 5 in each regime. First, we assume that the delay parameter is known and fix p max =p2 max =5 for =3, 5 and 0, so that taking into account that the number of possible values of the threshold parameter is /2, we compare 375, 625 and 250 models, respectively. In every case, we consider the following measures of the performance of the model selection criteria: a the frequency detection of the correct order p,p 2 =, ; b the root mean square error RMSE of estimation of the threshold parameter and c the root mean square prediction error RMSPE for the last observation by using the model chosen by each criteria, the fitted parameters and the true value. he results are in able. It can be seen that for small sample size, =30, the improvement in the number of times in which the correct model is selected can be as large as 30.5% see C and C in M, for =50 as large as 50.5% see AIC and AIC in M and for = 00 as large as 22.6% see AIC and AIC in M. First part of able 3 includes the improvement percentage of the number of times in which the correct orders are selected by all the modified criteria. We note that theaicu, AICu, Cu and Cu have larger frequency detection for =30 but the frequency detection decreases when the sample size increases. On the other hand, the RMSE of estimation of the threshold parameter are very close for the original and modified criteria, whereas the RMSPE is usually smaller for the modified criteria. Now, we assume that the delay is unknown and apply the same design as before with p max = p2 max = 5, d max = 4 and = 3, 5 and 0. Now we compare 500, 2500 and 5000 models, respectively. In every case, we consider the same three measures of the performance of the model selection criteria, the frequency detection of the correct order p,p 2, RMSE and RMSPE but we also include the frequency detection of selecting the correct delay parameter d and the frequency detection of the correct orders and delay parameter, p,p 2,d. he results are given in able 2.

10 P. Galeano, D. Peña / Journal of Statistical Planning and Inference able 2 Frequency of times of correct selection, root mean square errors of the threshold parameter and root mean square prediction errors assuming that d is unknown M = 30 BIC BIC AIC AIC AICc AICc AICu AICu C C Cc Cc Cu Cu p,p d p,p 2,d RMSE RMSPE p,p d p,p 2,d RMSE RMSPE = 50 p,p d p,p 2,d RMSE RMSPE p,p d p,p 2,d RMSE RMSPE = 00 p,p d p,p 2,d RMSE MSPE p,p d p,p 2,d RMSE RMSPE It can be seen that for small sample size, = 30, the improvement in the number of times in which the correct orders p,p 2,d=,, are selected can be as large as 43.8% see AIC and AIC in M2, for =50 as large as 74.2% see AIC and AIC in M2 and for =00 as large as 42.2% seeaic and AIC in M. See the second part of able 3 to find the improvement percentage of the number of times in which the correct orders and delay parameter are selected by the modified criteria. As in the case in which d is assumed known, the AICu, AICu, Cu and Cu have the larger frequency detection for the true autoregression orders and delay parameters for = 30, but the frequency detection decreases when the sample size increases. We note that sometimes the modified criteria have a shorter frequency detection of the delay parameter, but this is not a drawback for them because the aim is to detect both the true autoregressive orders and the delay parameter, and not only the delay parameter. Regarding the RMSE and the RMSPE, the results are similar to the case in which d is assumed to be known. In terms of computational effort, the time needed to select the model by all the criteria considered in this paper for a series with = 3, 5 and 0, when both the threshold and the delay parameter are unknown are 2.8, 8.9 and 547 s, respectively, using a program written in Matlab with a Pentium M.

11 282 P. Galeano, D. Peña / Journal of Statistical Planning and Inference able 3 Improvement percentage of the number of times in which the correct orders are selected by the modified criteria up, assuming d is known, down, assuming d is unknown M = 30 BIC AIC AICc AICu C Cc Cu d Known p,p p,p = 50 p,p p,p = 00 p,p p,p d Unknown = 30 p,p 2,d p,p 2,d = 50 p,p 2,d p,p 2,d = 00 p,p 2,d p,p 2,d Acknowledgments We would like to thank the associate editor and two anonymous referees for their helpful comments, and Jan De Gooijer for making his code available to us. We acknowledge financial support by Ministerio de Educación y Ciencia Grant SEJ and Comunidad de Madrid Grant HSE/074/2004. he first author also acknowledges financial support by Xunta de Galicia under the Isidro Parga Pondal Program. Appendix A. Proof of heorem. Shibata 980 considers order selection criteria of the form: S o p = p max + δ p + 2p σ 2 p. he order chosen for the selection criteria S o p is efficient if δ p verifies the conditions imposed in heorem 4.2 of Shibata 980 δ p. p lim max = 0, p p max p max δ p δ p 2. p lim max = 0, p p max p max L p where p lim denotes limit in probability, L p, is the following function: L p = pσ2 a p max + i=p+ j=p+ i j Σ ij,

12 P. Galeano, D. Peña / Journal of Statistical Planning and Inference where Σ ij = Covx t,x t i j and p is a sequence of positive integers with p p max which attain the minimum of L p for each see Shibata, 980, p. 54. he AIC can be written in terms of S o p taking δ p = δ AIC p = exp 2p p max 2p. Shibata 980 has shown that this term verifies the two conditions, and this gives the asymptotic efficiency of AIC. We can write AIC in terms of S o p taking δ p = δ AIC p = exp 2p log Q β p / p max 2p. herefore, 2p p = δ AIC p exp δ AIC log Q β p /. We show that δ AIC p verifies both conditions. First we write, δ AIC p exp 2p log Q β p / 2p = p max p max p max. 7 Hannan 973 shows that log Qγ /, for every γ belonging to the parametric space, and consequently, log Q β p / and the limit when of the maximum of the values 7 in the set p p max is 0. his proves the first condition. For the second condition, we write the following decomposition: δ AIC p δ AIC p p max L p δaic p δ AIC p p max L p 2p exp log Q β p / exp 2p log Q β p / +. p max L p Shibata 980 showed that the first term tends to 0 implying that AIC is efficient. For the second expression, for any p such that p p max including p, it can be shown that, 2p p lim exp log Q β p / = log i log 2 ii β p <. As this limit is bounded for every p and p max L p when, for every p p max, the second expression also tends to 0. hen, δ AIC p verifies the second condition. herefore, AIC is efficient. As AICc is asymptotically equivalent to AIC, AICc is also efficient. Proof of heorem 2. he BIC can be written as follows by using one step ahead prediction variances: BIC p = log σ 2 p + log vt 2 p + p + log = log σ 2 p + p + p + log vt 2 p + log. Now, note that the last term in the previous expression verifies: log vt 2 p + p + log, log vt 2 p + p + log 0. his shows that the criterion BIC is under the conditions of heorem 3 in Hannan 980, p. 073, implying that BIC is consistent.

13 284 P. Galeano, D. Peña / Journal of Statistical Planning and Inference References Akaike, H., 969. Fitting autoregressive models for prediction. Ann. Inst. Statist. Math. 2, Akaike, H., 973. Information theory and an extension of the maximum likelihood principle. In: Proceedings of the Second International Symposium on Information heory. Akademiai Kiadó, Budapest, pp Campbell, E.P., Bayesian selection of threshold autoregressive models. J. ime Ser. Anal. 25, Chan, K.S., 993. Consistency and limiting distribution of the least squares estimator of a threshold autoregressive model. Ann. Statist. 2, De Gooijer, J.G., 200. Cross-validation criteria for SEAR model selection. J. ime Ser. Anal. 22, Durbin, J., 960. he fitting of time series models. Rev. Internat. Statist. Inst. 28, Hannan, E.J., 973. he asymptotic theory of linear time-series models. J. Appl. Probab. 0, Hannan, E.J., 980. Estimation of the order of an ARMA process. Ann. Statist. 8, Hannan, E.J., Quinn, B.G., 979. he determination of the order of an autoregression. J. Roy. Statist. Soc. Ser. B 4, Harvey, A.C., 98. Forecasting, Structural ime Series Models and the Kalman Filter. Cambridge University Press, Cambridge. Hurvich, C.M., sai, C., 989. Regression and time series model selection in small samples. Biometrika 76, Hurvich, C.M., Shumway, R., sai, C., 990. Improved estimators of Kullback-Leibler information for autoregressive model selection in small samples. Biometrika 77, Kapetanios, G., 200. Model selection in threshold models. J. ime Ser. Anal. 22, van der Leeuw, J., 994. he covariance matrix of ARMA errors in closed form. J. Econometrics 63, McQuarrie, A., Shumway, R., sai, C.L., 997. he model selection criterion AICu. Statist. Probab. Lett. 34, Öhrvik, J., Schoier, G., SEAR model selection a bootstrap approach. Comput. Statist. 20, Peña, D., Rodriguez, J., he log of the determinant of the autocorrelation matrix for testing goodness of fit in time series. J. Statist. Plann. Inference 36, Ramsey, F.L., 974. Characterization of the partial autocorrelation function. Ann. Statist. 2, Schwarz, G., 978. Estimating the dimension of a model. Ann. Statist. 6, Shibata, R., 980. Asymptotically efficient selection of the order of the model for estimating parameters of a linear process. Ann. Statist. 8, Stoica, P., Eykhoff, P., Janssen, P., Söderström,., 986. Model-structure selection by cross-validation. Internat. J. Control 43, ong, H., 990. Non-linear ime Series: A Dynamical System Approach. Oxford University Press, Oxford. Unnikrishnan, N.K., Bayesian subset model selection for time series. J. ime Ser. Anal. 25, Wong, C.S., Li, W.K., 998. A note on the corrected Akaike information criterion for the threshold autoregressive models. J. ime Ser. Anal. 9, 3 24.

Outlier detection in ARIMA and seasonal ARIMA models by. Bayesian Information Type Criteria

Outlier detection in ARIMA and seasonal ARIMA models by. Bayesian Information Type Criteria Outlier detection in ARIMA and seasonal ARIMA models by Bayesian Information Type Criteria Pedro Galeano and Daniel Peña Departamento de Estadística Universidad Carlos III de Madrid 1 Introduction The

More information

Additive Outlier Detection in Seasonal ARIMA Models by a Modified Bayesian Information Criterion

Additive Outlier Detection in Seasonal ARIMA Models by a Modified Bayesian Information Criterion 13 Additive Outlier Detection in Seasonal ARIMA Models by a Modified Bayesian Information Criterion Pedro Galeano and Daniel Peña CONTENTS 13.1 Introduction... 317 13.2 Formulation of the Outlier Detection

More information

Performance of Autoregressive Order Selection Criteria: A Simulation Study

Performance of Autoregressive Order Selection Criteria: A Simulation Study Pertanika J. Sci. & Technol. 6 (2): 7-76 (2008) ISSN: 028-7680 Universiti Putra Malaysia Press Performance of Autoregressive Order Selection Criteria: A Simulation Study Venus Khim-Sen Liew, Mahendran

More information

1. Introduction Over the last three decades a number of model selection criteria have been proposed, including AIC (Akaike, 1973), AICC (Hurvich & Tsa

1. Introduction Over the last three decades a number of model selection criteria have been proposed, including AIC (Akaike, 1973), AICC (Hurvich & Tsa On the Use of Marginal Likelihood in Model Selection Peide Shi Department of Probability and Statistics Peking University, Beijing 100871 P. R. China Chih-Ling Tsai Graduate School of Management University

More information

On Autoregressive Order Selection Criteria

On Autoregressive Order Selection Criteria On Autoregressive Order Selection Criteria Venus Khim-Sen Liew Faculty of Economics and Management, Universiti Putra Malaysia, 43400 UPM, Serdang, Malaysia This version: 1 March 2004. Abstract This study

More information

Lecture 7: Model Building Bus 41910, Time Series Analysis, Mr. R. Tsay

Lecture 7: Model Building Bus 41910, Time Series Analysis, Mr. R. Tsay Lecture 7: Model Building Bus 41910, Time Series Analysis, Mr R Tsay An effective procedure for building empirical time series models is the Box-Jenkins approach, which consists of three stages: model

More information

Minimum Message Length Autoregressive Model Order Selection

Minimum Message Length Autoregressive Model Order Selection Minimum Message Length Autoregressive Model Order Selection Leigh J. Fitzgibbon School of Computer Science and Software Engineering, Monash University Clayton, Victoria 38, Australia leighf@csse.monash.edu.au

More information

Selection Criteria Based on Monte Carlo Simulation and Cross Validation in Mixed Models

Selection Criteria Based on Monte Carlo Simulation and Cross Validation in Mixed Models Selection Criteria Based on Monte Carlo Simulation and Cross Validation in Mixed Models Junfeng Shang Bowling Green State University, USA Abstract In the mixed modeling framework, Monte Carlo simulation

More information

The Influence of Additive Outliers on the Performance of Information Criteria to Detect Nonlinearity

The Influence of Additive Outliers on the Performance of Information Criteria to Detect Nonlinearity The Influence of Additive Outliers on the Performance of Information Criteria to Detect Nonlinearity Saskia Rinke 1 Leibniz University Hannover Abstract In this paper the performance of information criteria

More information

A TIME SERIES PARADOX: UNIT ROOT TESTS PERFORM POORLY WHEN DATA ARE COINTEGRATED

A TIME SERIES PARADOX: UNIT ROOT TESTS PERFORM POORLY WHEN DATA ARE COINTEGRATED A TIME SERIES PARADOX: UNIT ROOT TESTS PERFORM POORLY WHEN DATA ARE COINTEGRATED by W. Robert Reed Department of Economics and Finance University of Canterbury, New Zealand Email: bob.reed@canterbury.ac.nz

More information

Comparison of New Approach Criteria for Estimating the Order of Autoregressive Process

Comparison of New Approach Criteria for Estimating the Order of Autoregressive Process IOSR Journal of Mathematics (IOSRJM) ISSN: 78-578 Volume 1, Issue 3 (July-Aug 1), PP 1- Comparison of New Approach Criteria for Estimating the Order of Autoregressive Process Salie Ayalew 1 M.Chitti Babu,

More information

On the Central Limit Theorem for an ergodic Markov chain

On the Central Limit Theorem for an ergodic Markov chain Stochastic Processes and their Applications 47 ( 1993) 113-117 North-Holland 113 On the Central Limit Theorem for an ergodic Markov chain K.S. Chan Department of Statistics and Actuarial Science, The University

More information

SGN Advanced Signal Processing: Lecture 8 Parameter estimation for AR and MA models. Model order selection

SGN Advanced Signal Processing: Lecture 8 Parameter estimation for AR and MA models. Model order selection SG 21006 Advanced Signal Processing: Lecture 8 Parameter estimation for AR and MA models. Model order selection Ioan Tabus Department of Signal Processing Tampere University of Technology Finland 1 / 28

More information

Goodness-of-Fit Tests for Time Series Models: A Score-Marked Empirical Process Approach

Goodness-of-Fit Tests for Time Series Models: A Score-Marked Empirical Process Approach Goodness-of-Fit Tests for Time Series Models: A Score-Marked Empirical Process Approach By Shiqing Ling Department of Mathematics Hong Kong University of Science and Technology Let {y t : t = 0, ±1, ±2,

More information

9. Model Selection. statistical models. overview of model selection. information criteria. goodness-of-fit measures

9. Model Selection. statistical models. overview of model selection. information criteria. goodness-of-fit measures FE661 - Statistical Methods for Financial Engineering 9. Model Selection Jitkomut Songsiri statistical models overview of model selection information criteria goodness-of-fit measures 9-1 Statistical models

More information

Model selection criteria Λ

Model selection criteria Λ Model selection criteria Λ Jean-Marie Dufour y Université de Montréal First version: March 1991 Revised: July 1998 This version: April 7, 2002 Compiled: April 7, 2002, 4:10pm Λ This work was supported

More information

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY Time Series Analysis James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY PREFACE xiii 1 Difference Equations 1.1. First-Order Difference Equations 1 1.2. pth-order Difference Equations 7

More information

7. Forecasting with ARIMA models

7. Forecasting with ARIMA models 7. Forecasting with ARIMA models 309 Outline: Introduction The prediction equation of an ARIMA model Interpreting the predictions Variance of the predictions Forecast updating Measuring predictability

More information

Terence Tai-Leung Chong. Abstract

Terence Tai-Leung Chong. Abstract Estimation of the Autoregressive Order in the Presence of Measurement Errors Terence Tai-Leung Chong The Chinese University of Hong Kong Yuanxiu Zhang University of British Columbia Venus Liew Universiti

More information

10. Time series regression and forecasting

10. Time series regression and forecasting 10. Time series regression and forecasting Key feature of this section: Analysis of data on a single entity observed at multiple points in time (time series data) Typical research questions: What is the

More information

Asymptotic inference for a nonstationary double ar(1) model

Asymptotic inference for a nonstationary double ar(1) model Asymptotic inference for a nonstationary double ar() model By SHIQING LING and DONG LI Department of Mathematics, Hong Kong University of Science and Technology, Hong Kong maling@ust.hk malidong@ust.hk

More information

Variable Selection in Predictive Regressions

Variable Selection in Predictive Regressions Variable Selection in Predictive Regressions Alessandro Stringhi Advanced Financial Econometrics III Winter/Spring 2018 Overview This chapter considers linear models for explaining a scalar variable when

More information

Econ 623 Econometrics II Topic 2: Stationary Time Series

Econ 623 Econometrics II Topic 2: Stationary Time Series 1 Introduction Econ 623 Econometrics II Topic 2: Stationary Time Series In the regression model we can model the error term as an autoregression AR(1) process. That is, we can use the past value of the

More information

Econometrics I, Estimation

Econometrics I, Estimation Econometrics I, Estimation Department of Economics Stanford University September, 2008 Part I Parameter, Estimator, Estimate A parametric is a feature of the population. An estimator is a function of the

More information

Order Selection for Vector Autoregressive Models

Order Selection for Vector Autoregressive Models IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 51, NO. 2, FEBRUARY 2003 427 Order Selection for Vector Autoregressive Models Stijn de Waele and Piet M. T. Broersen Abstract Order-selection criteria for vector

More information

STAT Financial Time Series

STAT Financial Time Series STAT 6104 - Financial Time Series Chapter 4 - Estimation in the time Domain Chun Yip Yau (CUHK) STAT 6104:Financial Time Series 1 / 46 Agenda 1 Introduction 2 Moment Estimates 3 Autoregressive Models (AR

More information

[1] Thavaneswaran, A.; Heyde, C. C. A note on filtering for long memory processes. Stable non-gaussian models in finance and econometrics. Math.

[1] Thavaneswaran, A.; Heyde, C. C. A note on filtering for long memory processes. Stable non-gaussian models in finance and econometrics. Math. [1] Thavaneswaran, A.; Heyde, C. C. A note on filtering for long memory processes. Stable non-gaussian models in finance and econometrics. Math. Comput. Modelling 34 (2001), no. 9-11, 1139--1144. [2] Peiris,

More information

Model comparison and selection

Model comparison and selection BS2 Statistical Inference, Lectures 9 and 10, Hilary Term 2008 March 2, 2008 Hypothesis testing Consider two alternative models M 1 = {f (x; θ), θ Θ 1 } and M 2 = {f (x; θ), θ Θ 2 } for a sample (X = x)

More information

Performance of lag length selection criteria in three different situations

Performance of lag length selection criteria in three different situations MPRA Munich Personal RePEc Archive Performance of lag length selection criteria in three different situations Zahid Asghar and Irum Abid Quaid-i-Azam University, Islamabad Aril 2007 Online at htts://mra.ub.uni-muenchen.de/40042/

More information

Review Session: Econometrics - CLEFIN (20192)

Review Session: Econometrics - CLEFIN (20192) Review Session: Econometrics - CLEFIN (20192) Part II: Univariate time series analysis Daniele Bianchi March 20, 2013 Fundamentals Stationarity A time series is a sequence of random variables x t, t =

More information

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY Time Series Analysis James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY & Contents PREFACE xiii 1 1.1. 1.2. Difference Equations First-Order Difference Equations 1 /?th-order Difference

More information

Heteroskedasticity; Step Changes; VARMA models; Likelihood ratio test statistic; Cusum statistic.

Heteroskedasticity; Step Changes; VARMA models; Likelihood ratio test statistic; Cusum statistic. 47 3!,57 Statistics and Econometrics Series 5 Febrary 24 Departamento de Estadística y Econometría Universidad Carlos III de Madrid Calle Madrid, 126 2893 Getafe (Spain) Fax (34) 91 624-98-49 VARIANCE

More information

MASM22/FMSN30: Linear and Logistic Regression, 7.5 hp FMSN40:... with Data Gathering, 9 hp

MASM22/FMSN30: Linear and Logistic Regression, 7.5 hp FMSN40:... with Data Gathering, 9 hp Selection criteria Example Methods MASM22/FMSN30: Linear and Logistic Regression, 7.5 hp FMSN40:... with Data Gathering, 9 hp Lecture 5, spring 2018 Model selection tools Mathematical Statistics / Centre

More information

Forecasting Levels of log Variables in Vector Autoregressions

Forecasting Levels of log Variables in Vector Autoregressions September 24, 200 Forecasting Levels of log Variables in Vector Autoregressions Gunnar Bårdsen Department of Economics, Dragvoll, NTNU, N-749 Trondheim, NORWAY email: gunnar.bardsen@svt.ntnu.no Helmut

More information

Akaike Information Criterion

Akaike Information Criterion Akaike Information Criterion Shuhua Hu Center for Research in Scientific Computation North Carolina State University Raleigh, NC February 7, 2012-1- background Background Model statistical model: Y j =

More information

Vector Autoregressive Model. Vector Autoregressions II. Estimation of Vector Autoregressions II. Estimation of Vector Autoregressions I.

Vector Autoregressive Model. Vector Autoregressions II. Estimation of Vector Autoregressions II. Estimation of Vector Autoregressions I. Vector Autoregressive Model Vector Autoregressions II Empirical Macroeconomics - Lect 2 Dr. Ana Beatriz Galvao Queen Mary University of London January 2012 A VAR(p) model of the m 1 vector of time series

More information

Brief Sketch of Solutions: Tutorial 3. 3) unit root tests

Brief Sketch of Solutions: Tutorial 3. 3) unit root tests Brief Sketch of Solutions: Tutorial 3 3) unit root tests.5.4.4.3.3.2.2.1.1.. -.1 -.1 -.2 -.2 -.3 -.3 -.4 -.4 21 22 23 24 25 26 -.5 21 22 23 24 25 26.8.2.4. -.4 - -.8 - - -.12 21 22 23 24 25 26 -.2 21 22

More information

Comment about AR spectral estimation Usually an estimate is produced by computing the AR theoretical spectrum at (ˆφ, ˆσ 2 ). With our Monte Carlo

Comment about AR spectral estimation Usually an estimate is produced by computing the AR theoretical spectrum at (ˆφ, ˆσ 2 ). With our Monte Carlo Comment aout AR spectral estimation Usually an estimate is produced y computing the AR theoretical spectrum at (ˆφ, ˆσ 2 ). With our Monte Carlo simulation approach, for every draw (φ,σ 2 ), we can compute

More information

Bootstrap prediction and Bayesian prediction under misspecified models

Bootstrap prediction and Bayesian prediction under misspecified models Bernoulli 11(4), 2005, 747 758 Bootstrap prediction and Bayesian prediction under misspecified models TADAYOSHI FUSHIKI Institute of Statistical Mathematics, 4-6-7 Minami-Azabu, Minato-ku, Tokyo 106-8569,

More information

Econometric Forecasting

Econometric Forecasting Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna October 1, 2014 Outline Introduction Model-free extrapolation Univariate time-series models Trend

More information

Expressions for the covariance matrix of covariance data

Expressions for the covariance matrix of covariance data Expressions for the covariance matrix of covariance data Torsten Söderström Division of Systems and Control, Department of Information Technology, Uppsala University, P O Box 337, SE-7505 Uppsala, Sweden

More information

TIME SERIES ANALYSIS AND FORECASTING USING THE STATISTICAL MODEL ARIMA

TIME SERIES ANALYSIS AND FORECASTING USING THE STATISTICAL MODEL ARIMA CHAPTER 6 TIME SERIES ANALYSIS AND FORECASTING USING THE STATISTICAL MODEL ARIMA 6.1. Introduction A time series is a sequence of observations ordered in time. A basic assumption in the time series analysis

More information

Exercises - Time series analysis

Exercises - Time series analysis Descriptive analysis of a time series (1) Estimate the trend of the series of gasoline consumption in Spain using a straight line in the period from 1945 to 1995 and generate forecasts for 24 months. Compare

More information

ARIMA Modelling and Forecasting

ARIMA Modelling and Forecasting ARIMA Modelling and Forecasting Economic time series often appear nonstationary, because of trends, seasonal patterns, cycles, etc. However, the differences may appear stationary. Δx t x t x t 1 (first

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear

More information

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo Group Prof. Daniel Cremers 10a. Markov Chain Monte Carlo Markov Chain Monte Carlo In high-dimensional spaces, rejection sampling and importance sampling are very inefficient An alternative is Markov Chain

More information

Model selection using penalty function criteria

Model selection using penalty function criteria Model selection using penalty function criteria Laimonis Kavalieris University of Otago Dunedin, New Zealand Econometrics, Time Series Analysis, and Systems Theory Wien, June 18 20 Outline Classes of models.

More information

If we want to analyze experimental or simulated data we might encounter the following tasks:

If we want to analyze experimental or simulated data we might encounter the following tasks: Chapter 1 Introduction If we want to analyze experimental or simulated data we might encounter the following tasks: Characterization of the source of the signal and diagnosis Studying dependencies Prediction

More information

Thomas J. Fisher. Research Statement. Preliminary Results

Thomas J. Fisher. Research Statement. Preliminary Results Thomas J. Fisher Research Statement Preliminary Results Many applications of modern statistics involve a large number of measurements and can be considered in a linear algebra framework. In many of these

More information

CALCULATION METHOD FOR NONLINEAR DYNAMIC LEAST-ABSOLUTE DEVIATIONS ESTIMATOR

CALCULATION METHOD FOR NONLINEAR DYNAMIC LEAST-ABSOLUTE DEVIATIONS ESTIMATOR J. Japan Statist. Soc. Vol. 3 No. 200 39 5 CALCULAION MEHOD FOR NONLINEAR DYNAMIC LEAS-ABSOLUE DEVIAIONS ESIMAOR Kohtaro Hitomi * and Masato Kagihara ** In a nonlinear dynamic model, the consistency and

More information

Algorithm 808: ARfit A Matlab package for the estimation of parameters and eigenmodes of multivariate autoregressive models

Algorithm 808: ARfit A Matlab package for the estimation of parameters and eigenmodes of multivariate autoregressive models Algorithm 808: ARfit A Matlab package for the estimation of parameters and eigenmodes of multivariate autoregressive models TAPIO SCHNEIDER New York University and ARNOLD NEUMAIER Universität Wien ARfit

More information

Multivariate Time Series Analysis and Its Applications [Tsay (2005), chapter 8]

Multivariate Time Series Analysis and Its Applications [Tsay (2005), chapter 8] 1 Multivariate Time Series Analysis and Its Applications [Tsay (2005), chapter 8] Insights: Price movements in one market can spread easily and instantly to another market [economic globalization and internet

More information

Residuals in Time Series Models

Residuals in Time Series Models Residuals in Time Series Models José Alberto Mauricio Universidad Complutense de Madrid, Facultad de Económicas, Campus de Somosaguas, 83 Madrid, Spain. (E-mail: jamauri@ccee.ucm.es.) Summary: Three types

More information

Regression and Time Series Model Selection in Small Samples. Clifford M. Hurvich; Chih-Ling Tsai

Regression and Time Series Model Selection in Small Samples. Clifford M. Hurvich; Chih-Ling Tsai Regression and Time Series Model Selection in Small Samples Clifford M. Hurvich; Chih-Ling Tsai Biometrika, Vol. 76, No. 2. (Jun., 1989), pp. 297-307. Stable URL: http://links.jstor.org/sici?sici=0006-3444%28198906%2976%3a2%3c297%3aratsms%3e2.0.co%3b2-4

More information

Simulating Properties of the Likelihood Ratio Test for a Unit Root in an Explosive Second Order Autoregression

Simulating Properties of the Likelihood Ratio Test for a Unit Root in an Explosive Second Order Autoregression Simulating Properties of the Likelihood Ratio est for a Unit Root in an Explosive Second Order Autoregression Bent Nielsen Nuffield College, University of Oxford J James Reade St Cross College, University

More information

ARMA MODELS Herman J. Bierens Pennsylvania State University February 23, 2009

ARMA MODELS Herman J. Bierens Pennsylvania State University February 23, 2009 1. Introduction Given a covariance stationary process µ ' E[ ], the Wold decomposition states that where U t ARMA MODELS Herman J. Bierens Pennsylvania State University February 23, 2009 with vanishing

More information

Discrepancy-Based Model Selection Criteria Using Cross Validation

Discrepancy-Based Model Selection Criteria Using Cross Validation 33 Discrepancy-Based Model Selection Criteria Using Cross Validation Joseph E. Cavanaugh, Simon L. Davies, and Andrew A. Neath Department of Biostatistics, The University of Iowa Pfizer Global Research

More information

BOOTSTRAP PREDICTION INTERVALS IN STATE SPACE MODELS. Alejandro Rodriguez 1 and Esther Ruiz 2

BOOTSTRAP PREDICTION INTERVALS IN STATE SPACE MODELS. Alejandro Rodriguez 1 and Esther Ruiz 2 Working Paper 08-11 Departamento de Estadística Statistic and Econometric Series 04 Universidad Carlos III de Madrid March 2008 Calle Madrid, 126 28903 Getafe (Spain) Fax (34-91) 6249849 BOOTSTRAP PREDICTION

More information

Weighted tests of homogeneity for testing the number of components in a mixture

Weighted tests of homogeneity for testing the number of components in a mixture Computational Statistics & Data Analysis 41 (2003) 367 378 www.elsevier.com/locate/csda Weighted tests of homogeneity for testing the number of components in a mixture Edward Susko Department of Mathematics

More information

LATVIAN GDP: TIME SERIES FORECASTING USING VECTOR AUTO REGRESSION

LATVIAN GDP: TIME SERIES FORECASTING USING VECTOR AUTO REGRESSION LATVIAN GDP: TIME SERIES FORECASTING USING VECTOR AUTO REGRESSION BEZRUCKO Aleksandrs, (LV) Abstract: The target goal of this work is to develop a methodology of forecasting Latvian GDP using ARMA (AutoRegressive-Moving-Average)

More information

KALMAN-TYPE RECURSIONS FOR TIME-VARYING ARMA MODELS AND THEIR IMPLICATION FOR LEAST SQUARES PROCEDURE ANTONY G AU T I E R (LILLE)

KALMAN-TYPE RECURSIONS FOR TIME-VARYING ARMA MODELS AND THEIR IMPLICATION FOR LEAST SQUARES PROCEDURE ANTONY G AU T I E R (LILLE) PROBABILITY AND MATHEMATICAL STATISTICS Vol 29, Fasc 1 (29), pp 169 18 KALMAN-TYPE RECURSIONS FOR TIME-VARYING ARMA MODELS AND THEIR IMPLICATION FOR LEAST SQUARES PROCEDURE BY ANTONY G AU T I E R (LILLE)

More information

G. S. Maddala Kajal Lahiri. WILEY A John Wiley and Sons, Ltd., Publication

G. S. Maddala Kajal Lahiri. WILEY A John Wiley and Sons, Ltd., Publication G. S. Maddala Kajal Lahiri WILEY A John Wiley and Sons, Ltd., Publication TEMT Foreword Preface to the Fourth Edition xvii xix Part I Introduction and the Linear Regression Model 1 CHAPTER 1 What is Econometrics?

More information

Information Theory Based Estimator of the Number of Sources in a Sparse Linear Mixing Model

Information Theory Based Estimator of the Number of Sources in a Sparse Linear Mixing Model Information heory Based Estimator of the Number of Sources in a Sparse Linear Mixing Model Radu Balan University of Maryland Department of Mathematics, Center for Scientific Computation And Mathematical

More information

Department of Economics, Vanderbilt University While it is known that pseudo-out-of-sample methods are not optimal for

Department of Economics, Vanderbilt University While it is known that pseudo-out-of-sample methods are not optimal for Comment Atsushi Inoue Department of Economics, Vanderbilt University (atsushi.inoue@vanderbilt.edu) While it is known that pseudo-out-of-sample methods are not optimal for comparing models, they are nevertheless

More information

Volume 30, Issue 1. The relationship between the F-test and the Schwarz criterion: Implications for Granger-causality tests

Volume 30, Issue 1. The relationship between the F-test and the Schwarz criterion: Implications for Granger-causality tests Volume 30, Issue 1 The relationship between the F-test and the Schwarz criterion: Implications for Granger-causality tests Erdal Atukeren ETH Zurich - KOF Swiss Economic Institute Abstract In applied research,

More information

AR, MA and ARMA models

AR, MA and ARMA models AR, MA and AR by Hedibert Lopes P Based on Tsay s Analysis of Financial Time Series (3rd edition) P 1 Stationarity 2 3 4 5 6 7 P 8 9 10 11 Outline P Linear Time Series Analysis and Its Applications For

More information

Calculation of maximum entropy densities with application to income distribution

Calculation of maximum entropy densities with application to income distribution Journal of Econometrics 115 (2003) 347 354 www.elsevier.com/locate/econbase Calculation of maximum entropy densities with application to income distribution Ximing Wu Department of Agricultural and Resource

More information

The Role of "Leads" in the Dynamic Title of Cointegrating Regression Models. Author(s) Hayakawa, Kazuhiko; Kurozumi, Eiji

The Role of Leads in the Dynamic Title of Cointegrating Regression Models. Author(s) Hayakawa, Kazuhiko; Kurozumi, Eiji he Role of "Leads" in the Dynamic itle of Cointegrating Regression Models Author(s) Hayakawa, Kazuhiko; Kurozumi, Eiji Citation Issue 2006-12 Date ype echnical Report ext Version publisher URL http://hdl.handle.net/10086/13599

More information

Bias-corrected AIC for selecting variables in Poisson regression models

Bias-corrected AIC for selecting variables in Poisson regression models Bias-corrected AIC for selecting variables in Poisson regression models Ken-ichi Kamo (a), Hirokazu Yanagihara (b) and Kenichi Satoh (c) (a) Corresponding author: Department of Liberal Arts and Sciences,

More information

Akaike criterion: Kullback-Leibler discrepancy

Akaike criterion: Kullback-Leibler discrepancy Model choice. Akaike s criterion Akaike criterion: Kullback-Leibler discrepancy Given a family of probability densities {f ( ; ), 2 }, Kullback-Leibler s index of f ( ; ) relativetof ( ; ) is Z ( ) =E

More information

Standard Errors & Confidence Intervals. N(0, I( β) 1 ), I( β) = [ 2 l(β, φ; y) β i β β= β j

Standard Errors & Confidence Intervals. N(0, I( β) 1 ), I( β) = [ 2 l(β, φ; y) β i β β= β j Standard Errors & Confidence Intervals β β asy N(0, I( β) 1 ), where I( β) = [ 2 l(β, φ; y) ] β i β β= β j We can obtain asymptotic 100(1 α)% confidence intervals for β j using: β j ± Z 1 α/2 se( β j )

More information

ISSN Article. Selection Criteria in Regime Switching Conditional Volatility Models

ISSN Article. Selection Criteria in Regime Switching Conditional Volatility Models Econometrics 2015, 3, 289-316; doi:10.3390/econometrics3020289 OPEN ACCESS econometrics ISSN 2225-1146 www.mdpi.com/journal/econometrics Article Selection Criteria in Regime Switching Conditional Volatility

More information

End-Semester Examination MA 373 : Statistical Analysis on Financial Data

End-Semester Examination MA 373 : Statistical Analysis on Financial Data End-Semester Examination MA 373 : Statistical Analysis on Financial Data Instructor: Dr. Arabin Kumar Dey, Department of Mathematics, IIT Guwahati Note: Use the results in Section- III: Data Analysis using

More information

Saskia Rinke and Philipp Sibbertsen* Information criteria for nonlinear time series models

Saskia Rinke and Philipp Sibbertsen* Information criteria for nonlinear time series models Stud. Nonlinear Dyn. E. 16; (3): 35 341 Saskia Rinke and Philipp Sibbertsen* Information criteria for nonlinear time series models DOI 1.1515/snde-15-6 Abstract: In this paper the performance of different

More information

Model Selection for Semiparametric Bayesian Models with Application to Overdispersion

Model Selection for Semiparametric Bayesian Models with Application to Overdispersion Proceedings 59th ISI World Statistics Congress, 25-30 August 2013, Hong Kong (Session CPS020) p.3863 Model Selection for Semiparametric Bayesian Models with Application to Overdispersion Jinfang Wang and

More information

UNBIASED ESTIMATE FOR b-value OF MAGNITUDE FREQUENCY

UNBIASED ESTIMATE FOR b-value OF MAGNITUDE FREQUENCY J. Phys. Earth, 34, 187-194, 1986 UNBIASED ESTIMATE FOR b-value OF MAGNITUDE FREQUENCY Yosihiko OGATA* and Ken'ichiro YAMASHINA** * Institute of Statistical Mathematics, Tokyo, Japan ** Earthquake Research

More information

Lecture 6: Model Checking and Selection

Lecture 6: Model Checking and Selection Lecture 6: Model Checking and Selection Melih Kandemir melih.kandemir@iwr.uni-heidelberg.de May 27, 2014 Model selection We often have multiple modeling choices that are equally sensible: M 1,, M T. Which

More information

Departamento de Estadfstica y Econometrfa Statistics and Econometrics Series 27. Universidad Carlos III de Madrid December 1993 Calle Madrid, 126

Departamento de Estadfstica y Econometrfa Statistics and Econometrics Series 27. Universidad Carlos III de Madrid December 1993 Calle Madrid, 126 -------_._-- Working Paper 93-45 Departamento de Estadfstica y Econometrfa Statistics and Econometrics Series 27 Universidad Carlos III de Madrid December 1993 Calle Madrid, 126 28903 Getafe (Spain) Fax

More information

FORECASTING SUGARCANE PRODUCTION IN INDIA WITH ARIMA MODEL

FORECASTING SUGARCANE PRODUCTION IN INDIA WITH ARIMA MODEL FORECASTING SUGARCANE PRODUCTION IN INDIA WITH ARIMA MODEL B. N. MANDAL Abstract: Yearly sugarcane production data for the period of - to - of India were analyzed by time-series methods. Autocorrelation

More information

A Course in Time Series Analysis

A Course in Time Series Analysis A Course in Time Series Analysis Edited by DANIEL PENA Universidad Carlos III de Madrid GEORGE C. TIAO University of Chicago RUEY S. TSAY University of Chicago A Wiley-Interscience Publication JOHN WILEY

More information

Regression, Ridge Regression, Lasso

Regression, Ridge Regression, Lasso Regression, Ridge Regression, Lasso Fabio G. Cozman - fgcozman@usp.br October 2, 2018 A general definition Regression studies the relationship between a response variable Y and covariates X 1,..., X n.

More information

DYNAMIC ECONOMETRIC MODELS Vol. 9 Nicolaus Copernicus University Toruń Mariola Piłatowska Nicolaus Copernicus University in Toruń

DYNAMIC ECONOMETRIC MODELS Vol. 9 Nicolaus Copernicus University Toruń Mariola Piłatowska Nicolaus Copernicus University in Toruń DYNAMIC ECONOMETRIC MODELS Vol. 9 Nicolaus Copernicus University Toruń 2009 Mariola Piłatowska Nicolaus Copernicus University in Toruń Combined Forecasts Using the Akaike Weights A b s t r a c t. The focus

More information

1 Introduction In this work we pursue a new family of models for time series, substantially extending, but strongly related to and based on the classi

1 Introduction In this work we pursue a new family of models for time series, substantially extending, but strongly related to and based on the classi Time Series Prediction Using Mixtures of Experts Assaf J. Zeevi Information Systems Lab Department of Electrical Engineering Stanford University Stanford, CA. 94305 azeevi@isl.stanford.edu Ron Meir Department

More information

Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices

Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices Article Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices Fei Jin 1,2 and Lung-fei Lee 3, * 1 School of Economics, Shanghai University of Finance and Economics,

More information

Mohsen Pourahmadi. 1. A sampling theorem for multivariate stationary processes. J. of Multivariate Analysis, Vol. 13, No. 1 (1983),

Mohsen Pourahmadi. 1. A sampling theorem for multivariate stationary processes. J. of Multivariate Analysis, Vol. 13, No. 1 (1983), Mohsen Pourahmadi PUBLICATIONS Books and Editorial Activities: 1. Foundations of Time Series Analysis and Prediction Theory, John Wiley, 2001. 2. Computing Science and Statistics, 31, 2000, the Proceedings

More information

Lecture 2: Univariate Time Series

Lecture 2: Univariate Time Series Lecture 2: Univariate Time Series Analysis: Conditional and Unconditional Densities, Stationarity, ARMA Processes Prof. Massimo Guidolin 20192 Financial Econometrics Spring/Winter 2017 Overview Motivation:

More information

DEPARTMENT OF ECONOMETRICS AND BUSINESS STATISTICS

DEPARTMENT OF ECONOMETRICS AND BUSINESS STATISTICS ISSN 1440-771X AUSTRALIA DEPARTMENT OF ECONOMETRICS AND BUSINESS STATISTICS Empirical Information Criteria for Time Series Forecasting Model Selection Md B Billah, R.J. Hyndman and A.B. Koehler Working

More information

Statistics 203: Introduction to Regression and Analysis of Variance Course review

Statistics 203: Introduction to Regression and Analysis of Variance Course review Statistics 203: Introduction to Regression and Analysis of Variance Course review Jonathan Taylor - p. 1/?? Today Review / overview of what we learned. - p. 2/?? General themes in regression models Specifying

More information

Stationary Stochastic Time Series Models

Stationary Stochastic Time Series Models Stationary Stochastic Time Series Models When modeling time series it is useful to regard an observed time series, (x 1,x,..., x n ), as the realisation of a stochastic process. In general a stochastic

More information

Measures of Fit from AR(p)

Measures of Fit from AR(p) Measures of Fit from AR(p) Residual Sum of Squared Errors Residual Mean Squared Error Root MSE (Standard Error of Regression) R-squared R-bar-squared = = T t e t SSR 1 2 ˆ = = T t e t p T s 1 2 2 ˆ 1 1

More information

On Forecast Strength of Some Linear and Non Linear Time Series Models for Stationary Data Structure

On Forecast Strength of Some Linear and Non Linear Time Series Models for Stationary Data Structure American Journal of Mathematics and Statistics 2015, 5(4): 163-177 DOI: 10.5923/j.ajms.20150504.01 On Forecast Strength of Some Linear and Non Linear Imam Akeyede 1,*, Babatunde Lateef Adeleke 2, Waheed

More information

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley Time Series Models and Inference James L. Powell Department of Economics University of California, Berkeley Overview In contrast to the classical linear regression model, in which the components of the

More information

Bayesian Inference for DSGE Models. Lawrence J. Christiano

Bayesian Inference for DSGE Models. Lawrence J. Christiano Bayesian Inference for DSGE Models Lawrence J. Christiano Outline State space-observer form. convenient for model estimation and many other things. Bayesian inference Bayes rule. Monte Carlo integation.

More information

Math 423/533: The Main Theoretical Topics

Math 423/533: The Main Theoretical Topics Math 423/533: The Main Theoretical Topics Notation sample size n, data index i number of predictors, p (p = 2 for simple linear regression) y i : response for individual i x i = (x i1,..., x ip ) (1 p)

More information

Reduced rank regression in cointegrated models

Reduced rank regression in cointegrated models Journal of Econometrics 06 (2002) 203 26 www.elsevier.com/locate/econbase Reduced rank regression in cointegrated models.w. Anderson Department of Statistics, Stanford University, Stanford, CA 94305-4065,

More information

An EM algorithm for Gaussian Markov Random Fields

An EM algorithm for Gaussian Markov Random Fields An EM algorithm for Gaussian Markov Random Fields Will Penny, Wellcome Department of Imaging Neuroscience, University College, London WC1N 3BG. wpenny@fil.ion.ucl.ac.uk October 28, 2002 Abstract Lavine

More information

Bias and variance reduction techniques for bootstrap information criteria

Bias and variance reduction techniques for bootstrap information criteria Ann Inst Stat Math (2010) 62:209 234 DOI 10.1007/s10463-009-0237-1 Bias and variance reduction techniques for bootstrap information criteria Genshiro Kitagawa Sadanori Konishi Received: 17 November 2008

More information

Gaussian processes. Basic Properties VAG002-

Gaussian processes. Basic Properties VAG002- Gaussian processes The class of Gaussian processes is one of the most widely used families of stochastic processes for modeling dependent data observed over time, or space, or time and space. The popularity

More information

Topic 4 Unit Roots. Gerald P. Dwyer. February Clemson University

Topic 4 Unit Roots. Gerald P. Dwyer. February Clemson University Topic 4 Unit Roots Gerald P. Dwyer Clemson University February 2016 Outline 1 Unit Roots Introduction Trend and Difference Stationary Autocorrelations of Series That Have Deterministic or Stochastic Trends

More information

interval forecasting

interval forecasting Interval Forecasting Based on Chapter 7 of the Time Series Forecasting by Chatfield Econometric Forecasting, January 2008 Outline 1 2 3 4 5 Terminology Interval Forecasts Density Forecast Fan Chart Most

More information