A Comparison between Linear and Nonlinear Forecasts for Nonlinear AR Models

Similar documents
Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley

Notes on Time Series Modeling

2014 Preliminary Examination

ECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications

Joint Statistical Meetings - Section on Statistics & the Environment

Financial Econometrics

1 The Multiple Regression Model: Freeing Up the Classical Assumptions

2. Multivariate ARMA

Econ 423 Lecture Notes: Additional Topics in Time Series 1

The Predictive Space or If x predicts y; what does y tell us about x?

May 2, Why do nonlinear models provide poor. macroeconomic forecasts? Graham Elliott (UCSD) Gray Calhoun (Iowa State) Motivating Problem

Testing for Regime Switching: A Comment

data lam=36.9 lam=6.69 lam=4.18 lam=2.92 lam=2.21 time max wavelength modulus of max wavelength cycle

Bias Correction in the Balanced-half-sample Method if the Number of Sampled Units in Some Strata Is Odd

Measuring robustness

Estimating the Number of Common Factors in Serially Dependent Approximate Factor Models

THE GEOMETRICAL ERGODICITY OF NONLINEAR AUTOREGRESSIVE MODELS

Parametric Inference on Strong Dependence

Econometric Forecasting

TIME SERIES DATA ANALYSIS USING EVIEWS

G. S. Maddala Kajal Lahiri. WILEY A John Wiley and Sons, Ltd., Publication

Elements of Multivariate Time Series Analysis

Chapter 2. Dynamic panel data models

TIME SERIES MODELLING OF SURFACE PRESSURE DATA

ECONOMETRICS FIELD EXAM Michigan State University May 9, 2008

A Course in Time Series Analysis

GMM-based inference in the AR(1) panel data model for parameter values where local identi cation fails

Economics 241B Estimation with Instruments

On the Central Limit Theorem for an ergodic Markov chain

1 A Non-technical Introduction to Regression

Modeling Expectations with Noncausal Autoregressions

Production Policies for Multi-Product Systems with Deteriorating. Process Condition

The Kuhn-Tucker Problem

The Generalized Cochrane-Orcutt Transformation Estimation For Spurious and Fractional Spurious Regressions

Nonlinear Time Series Modeling

x i = 1 yi 2 = 55 with N = 30. Use the above sample information to answer all the following questions. Show explicitly all formulas and calculations.

Chapter 2. GMM: Estimating Rational Expectations Models

UNIVERSITY OF CALIFORNIA, SAN DIEGO DEPARTMENT OF ECONOMICS

Economics 620, Lecture 13: Time Series I

Simple Estimators for Monotone Index Models

NUMERICAL DISTRIBUTION FUNCTIONS OF LIKELIHOOD RATIO TESTS FOR COINTEGRATION

Technical Appendix-3-Regime asymmetric STAR modeling and exchange rate reversion

Modeling Expectations with Noncausal Autoregressions

Chapter 1. GMM: Basic Concepts

Online Appendix to: Marijuana on Main Street? Estimating Demand in Markets with Limited Access

Lecture 6: Univariate Volatility Modelling: ARCH and GARCH Models

Problem set 1 - Solutions

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

1. Introduction Over the last three decades a number of model selection criteria have been proposed, including AIC (Akaike, 1973), AICC (Hurvich & Tsa

SIMILAR-ON-THE-BOUNDARY TESTS FOR MOMENT INEQUALITIES EXIST, BUT HAVE POOR POWER. Donald W. K. Andrews. August 2011

Introduction to Eco n o m et rics

CS 540: Machine Learning Lecture 2: Review of Probability & Statistics

Supplemental Material 1 for On Optimal Inference in the Linear IV Model

Cointegration Tests Using Instrumental Variables Estimation and the Demand for Money in England

On Backtesting Risk Measurement Models

If we want to analyze experimental or simulated data we might encounter the following tasks:

A Conditional-Heteroskedasticity-Robust Con dence Interval for the Autoregressive Parameter

Simple Estimators for Semiparametric Multinomial Choice Models

ARIMA Models. Richard G. Pierse

Do Markov-Switching Models Capture Nonlinearities in the Data? Tests using Nonparametric Methods

Performance of Autoregressive Order Selection Criteria: A Simulation Study

Short-run electricity demand forecasts in Maharashtra

Time Series Analysis -- An Introduction -- AMS 586

Time Series: Theory and Methods

11. Bootstrap Methods

Statistics of Stochastic Processes

TAKEHOME FINAL EXAM e iω e 2iω e iω e 2iω

by Søren Johansen Department of Economics, University of Copenhagen and CREATES, Aarhus University

FORECASTING AND MODEL SELECTION

Forecasting the unemployment rate when the forecast loss function is asymmetric. Jing Tian

MC3: Econometric Theory and Methods. Course Notes 4

ARMA Models: I VIII 1

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

Chapter 6: Endogeneity and Instrumental Variables (IV) estimator

Nonlinear Time Series

Implicit Function Theorem: One Equation

Departamento de Estadística y Econometría. January 2001 Calle Madrid, Getafe (Spain) Fax (34-91)

Quadratic forms in skew normal variates

ECON0702: Mathematical Methods in Economics

13 Endogeneity and Nonparametric IV

Decision 411: Class 7

Research Article Least Squares Estimators for Unit Root Processes with Locally Stationary Disturbance

Common Core State Standards with California Additions 1 Standards Map. Algebra I

LINEAR MMSE ESTIMATION

Iterative Stepwise Selection and Threshold for Learning Causes in Time Series

E cient Regressions via Optimally Combining Quantile Information

GMM estimation of spatial panels

1 Regression with Time Series Variables

Economics 620, Lecture 18: Nonlinear Models

A CONDITIONAL-HETEROSKEDASTICITY-ROBUST CONFIDENCE INTERVAL FOR THE AUTOREGRESSIVE PARAMETER. Donald W.K. Andrews and Patrik Guggenberger

On Autoregressive Order Selection Criteria

Stochastic Processes: I. consider bowl of worms model for oscilloscope experiment:

THE MINIMAL OVERLAP COST SHARING RULE

MEASURING THE SOCIO-ECONOMIC BIPOLARIZATION PHENOMENON

Modeling and Performance Analysis with Discrete-Event Simulation

Lecture 8: ARIMA Forecasting Please read Chapters 7 and 8 of MWH Book

Nonparametric Identi cation and Estimation of Truncated Regression Models with Heteroskedasticity

LECTURE 13: TIME SERIES I

Efficiency Tradeoffs in Estimating the Linear Trend Plus Noise Model. Abstract

Can a subset of forecasters beat the simple average in the SPF?

Transcription:

JOURNAL OF FORECASTING, VOL. 6, 49±508 (997) A Comparison between Linear and Nonlinear Forecasts for Nonlinear AR Models MEIHUI GUO* AND Y. K. TSENG National Sun Yat-sen University, Taiwan, ROC ABSTRACT In this paper the relative forecast performance of nonlinear models to linear models is assessed by the conditional probability that the absolute forecast error of the nonlinear forecast is smaller than that of the linear forecast. The comparison probability is explicitly expressed and is shown to be an increasing function of the distance between nonlinear and linear forecasts under certain conditions. This expression of the comparison probability may not only be useful in determining the predictor, which is either a more accurate or a simpler forecast, to be used but also provides a good explanation for an odd phenomenon discussed by Pemberton. The relative forecast performance of a nonlinear model to a linear model is demonstrated to be sensitive to its forecast origins. A new forecast is thus proposed to improve the relative forecast performance of nonlinear models based on forecast origins. # 997 John Wiley & Sons, Ltd. J. forecast, Vol. 6, 49±508 (997) No. of Figures: 3. No. of Tables:. No. of References: 8. KEY WORDS forecast origin; multi-step forecast; SETAR model INTRODUCTION Prediction is an important topic in time series analysis. Many stationary phenomena in practice can be described or at least be approximated by stationary linear time series models, e.g. the ARMA( p, q) models (see, for example, Priestly, 989). However, many nonlinear phenomena such as limit cycles, frequency modulations (Tong, 990) and animal population cycles (Oster and Ipaktchi, 978) cannot be described adequately by linear time series models, unless super uous parameters are involved with. Intuitively, if the true model is a nonlinear time series model, then any statistical inferences using analysis for the nonlinear model, which capture the nonlinear characteristic of the data, should be better than those by using linear approximation. * Correspondence to: Meihui Guo, Department of Applied Mathematics, National Sun Yat-sen University, Kaohsiung, ROC 8044, Taiwan. Contract grant sponsor: National Science Council of the Republic of China; Contract grant number: NSC 8-008-N- 0-064. CCC 077±6693/97/07049±8$7.50 Received January 996 # 997 John Wiley & Sons, Ltd. Accepted July 997

49 Meihui Guo and Y. K. Tseng However, this is not always the case for the prediction problem. For example, Davies et al. (988) and Pemberton (989), by numerical simulation, observe the phenomenon that the conditional mean and conditional median forecasts of nonlinear time series models have poor forecast performance compared to those of linear models. Therefore, if the linear forecast is not very much worse than the nonlinear forecast, then it is reasonable to adopt the former forecast rather than the nonlinear forecast, at least for computational reasons. More importantly, in practical applications for a given data set, unless an additional test is done in advance, there is no way of knowing whether the true model is factually linear or nonlinear. Thus, there arises the question: Under what circumstances should the nonlinear predictors be adopted? The present research is motivated by Davies et al. (988) who conducted three experiments to compare the forecast performance of nonlinear SETAR (Self-Exciting Threshold AutoRegressive) models with linear AR models. In experiment I they carried out the identi cation and estimation of SETAR models for real time series data sets. In experiments II and III they simulated series for di erent lengths respectively from each of 50 randomly selected stationary SETAR(;,) models of the form X n ˆ ax n e n if X n 4 0 bx n e n if X n > 0 where e n 's are i.i.d. N(0, ) r.v's. For each set of the data, both nonlinear SETAR models and linear AR models were tted. Among three experiments the relative frequencies, of the events that the absolute forecast error of an m-step (m ˆ,, 3) nonlinear forecast is smaller than that of the linear AR forecast, range from 48% to 69% when the conditional mean is used as the forecast and range from 47% to 7% when the conditional median is used. These results strongly imply that there is a positive probability that the linear forecast is better than the nonlinear forecast even when the true model is nonlinear. Pemberton (989) conducted simulations for model () with a ˆ b ˆ 0.4 and 0.9 and tted generated data by AR models. For the conditional median forecast, the conditional probability that the absolute forecast error of the m-step nonlinear forecast is smaller than that of the AR forecast given ~X t ˆ (X,..., X t ), is ) ˆ max{f m (Q m ) j ~X t ), 7 F m (Q m ) j ~X t )}, where F m (. j ~X t ) is the conditional distribution function of X t m given ~X t and Q m ) is the midpoint of the two forecasts. It can be shown that ) is not less than /. Pemberton, by plotting ) against the values of the stationary density function of the true model, observed that the nonlinear conditional median forecast performs better at those forecast origin points whose probability density has low values than at those points whose probability density has high ones (in the sense that ) is far greater than or close to 0.5). Pemberton also conducted his simulation for the simple piecewise constant model: X n ˆ a e n if X n 4 0 a e n if X n > 0 with a ˆ 3 and tted the data with a linear AR model by least squares method. Pemberton observed that () for lags m ˆ,,..., 30, the minima of ), RMSE and RMAE appear in a vicinity of the modes of the stationary distribution. The RMSE (RMAE) denotes the Ratio of the Mean Squared (Absolute) Error of the conditional mean (median, correspondingly) forecast by using the nonlinear model to that by using the AR model; () when m!, RMSE (decreasingly) and RMAE (not necessarily decreasingly) tend to one whereas ) tends to /. J. forecast, Vol. 6, 49±508 (997) # 997 John Wiley & Sons, Ltd.

Nonlinear AR Models 493 Recently, Tiao and Tsay (994) compare the post-sample forecast of TAR model versus linear AR() model for US quarterly real GNP data by the ratio of mean squared forecast errors. The nonlinear TAR model only outperforms the linear AR model by a substantial margin in shortterm forecasts at two regimes of the four forecast regimes. All the above examples show that for the prediction problem, the nonlinear forecast may not be more accurate than the linear forecast, even when the true model is nonlinear and that both linear and nonlinear forecasts are asymptotically equivalent for long-term prediction. In order to better understand the poor forecast problem of the nonlinear time series models, the properties of the forecast performance criterion ) considered by both Davies et al. and Pemberton are studied in this paper. Under certain conditions, the criterion ) is shown to be an increasing function of the distance between linear and nonlinear forecasts. This result gives a nice explanation for the odd phenomena discussed by Pemberton (989). In their simulation studies, since the distance between two forecasts are minimized around the modes of the stationary distribution of the process, the criterion ) is minimized at forecast regions of high probability density (see Example below). For time series forecasting, usually either conditional means or conditional medians are used as optimal forecasts. However, from the above-mentioned studies and as will also be shown in this paper, the relative forecast performance (for both conditional mean and conditional median) of a nonlinear model to a linear model is rather sensitive to the forecast origins. Given a nonlinear time series, one can theoretically nd a linear model which well approximates the true nonlinear models. The organization of the paper is as follows. In the next section de nitions and symbols are presented. In the third section the relation between ) and the distance between linear and nonlinear forecasts is exposed. An example is given to illustrate the odd phenomenon discussed by Pemberton. In the fourth section a new method is suggested which selects forecasts from conditional means or conditional medians by comparing their forecast performance with linear models. An example of piecewise constant SETAR model is also given. General conclusions will be given in the fth section. All the technical proofs are given in the Appendix. DEFINITIONS AND NOTATIONS Consider the nonlinear autoregressive model of order p(nlar( p)) X n ˆ l 0 X n,..., X n p e n 3 where l 0 (. ) is a measurable function from R p to R, e n 's are i.i.d. r.v's with mean zero, variance s and a common probability density function g(. ). Model (3) includes linear, SETAR and exponentially autoregressive (see Ozaki, 985) models as special cases. Throughout this paper, the process {X n } de ned by model (3) is assumed to be stationary. Assuming the observations ~X t ˆ (X, X,..., X t ) (capitals are used for both random variables and realizations) is generated from model (3), the unknown value of X t m (m 5 ) will be predicted by either the conditional mean or the conditional median when ~X t is given. Since the conditional mean E m j ~X t ) minimizes the mean squared error, it is also called the MMSE forecast. Similarly, the conditional median of X t m when ~X t is given will be called the MMAE (Minimum Mean Absolute Error) forecast. In general, there are no closed forms of the MMSE and MMAE forecasts for NLAR( p) models, except some numerical algorithms that may be available. The # 997 John Wiley & Sons, Ltd. J. forecast, Vol. 6, 49±508 (997)

494 Meihui Guo and Y. K. Tseng goal of this paper is to compare the NLAR(p) forecast with the linear AR(p) forecast when the true model is factually NLAR( p). De ne ˆ P j X t m NL m j<j X t m AR m k; ~X t 4 where NL m ) is the nonlinear m-step MMSE or MMAE forecast and AR m ) is the linear AR(p) m-step conditional mean forecast. When NL m 6ˆ AR m, it can be shown that ˆ F m j ~X t I R Š F m j ~X t I R Š 5 where F m (. j ~X t ) is the conditional distribution of X t m given ~X t, Q m ˆ NL m AR m, I Š is the indicator function, R ˆ { ~X t : NL m ) < AR m )}, and R ˆ { ~X t : NL m ) > AR m )}. The probability ) is considered as a criterion to compare the forecast performance, by Davies et al. as well as by Pemberton. Finally, we denote x m ) ˆ j NL m ) 7 AR m j, the distance between the nonlinear and linear forecasts at the forecast origin ~X t. FORECAST ACCURACY AND FORECAST DISTANCE In this section the comparison probability ) and the ratio of mean squared forecast errors are shown to be increasing functions of the forecast distance x m ) (see Theorems and 5 below). Throughout this paper, m m ) denotes the m-step conditional mean and M m ) the m-step conditional median of the NLAR( p) model. Theorem Suppose that (X, X,..., X t ) is observed from the NLAR(p) model (3). If we write F m (x j ~X t ) ˆ F(x 7 NL m )), then ˆ F x m I R Š F x m I R Š 6 This equation shows that for the MMAE forecast, ) 5 / a.s., if P(x m ) ˆ 0) ˆ 0. If the distribution F(. ) is continuous and symmetric about 0, then ˆ F x m. Corollary Let G(. ) denote the distribution function of the innovations. For multi-step forecast (i.e. m > ), the distribution function F in Theorem is given by F(x) ˆ E[G(x 7 l 0 m7,..., X t m7p ) NL m )) j ~X t ]. Especially for the one-step MMSE forecast, since NL ) ˆ l 0,..., X t 7p ), we have P ˆ G x I R Š G x I R Š. The proofs of Theorem and Corollary are immediate and hence omitted. Note that ) is the probability that the nonlinear forecast is better than the linear forecast. Therefore, we may say that the nonlinear forecast is better than the linear forecast if > =. Theorem shows that the nonlinear MMAE forecast is not worse than the linear MMAE forecast. However, it should be noted that this conclusion is not true practically since the theoretic nonlinear forecast NL m ) is factually unknown in real situations. In real applications, the nonlinear AR function l 0 m7,..., X t m7p ) is usually unknown and only its estimates are available. Even if it were known, the forecast NL m ) cannot be exactly computed and only J. forecast, Vol. 6, 49±508 (997) # 997 John Wiley & Sons, Ltd.

Nonlinear AR Models 495 an approximation of it may be available. This shows that the nonlinear forecast may not be actually better than the linear forecast. For the MMSE forecasts, the nonlinear forecast may not be better than the linear forecast, even theoretically. For example, when the distribution of the innovations satis es G x < = and the origin is in the set R ; or G x < = and the origin is in the set R. If the error in the computation of the MMSE nonlinear forecast is taken into account, the MMSE nonlinear forecast will further lose its superiority to the linear forecast. If the regressor l 0 has an explicit form and the distribution of the innovations is known, sometimes the conditional distribution F m (x j ~X t ) of X t m can be explicitly expressed. As an example, consider the piecewise constant SETAR model: X n ˆ a j e n if X n c j ; c j Š 7 where j ˆ,,..., l, ˆ c 0 < c < < c l ˆ, e n 's are i.i.d. N(0, ) random variables. Theorem 3 Consider the m-step MMSE forecast when the true model is (7). In this case, we have NL m ) ˆ m m ). When X t c j, c j ], j ˆ,,..., l, we have x ˆ ˆ H m m;j I R Š H m;j x m I R Š where H m;j x ˆ Pl iˆ k m ij F x a m ij with K ˆ (F(c i 7 a j ) 7 F(c i7 7 a j )) and a m ij, 8j ˆ,,..., l, k m ij are elements of the matrix K m 7 ij a. ˆ a i P l iˆ k m Corollary 4 For model (), NL m ) ˆ m m ) ˆ ab m 7 sign ), the comparison probability has the following explicit form: x ˆ H m m I R Š H m x m I R Š where H m x ˆ k m F x sign ak m k m F x sign ak m, k m ˆ b m =, k m ˆ b m =, and b ˆ 7 F(a). Now, consider the ratio of the mean squared (absolute) errors of the nonlinear forecasts to linear forecasts under model (3). De ne where and de ne RMSE m ˆ m AR m j ~X t Š MSE m m j ~X t MSE m m j ~X t ˆ E m m m j ~X t Š ˆ s E l 0 m ;... ; X t m p m m j ~X t Š RMAE m ˆ E j m AR m k ~X t Š MAE M m j ~X t where MAE(M m ) j ~X t ) ˆ E[j m 7 M m ) k ~X t ]. # 997 John Wiley & Sons, Ltd. J. forecast, Vol. 6, 49±508 (997)

496 Meihui Guo and Y. K. Tseng The following theorem shows that both RMSE m ) and RMAE ) are increasing functions of x m ). Theorem 5 If model (3) is the true model, then and RMAE m ˆ x RMSE m ˆ m MSE m m j ~X t j R AR m x AR M m m df m x j ~X t j MAE M m j ~X t Next, let us consider the m-step forecast for large m when model (3) is the true model. We say that the time series {X t } de ned by model (3) is causal if for all t, X t can be written as a function of e t, e t7,.... In this case, we have s ; X t ;... s e t ; e t ;.... When {X t } de ned by model (3) is stationary and causal, we have m m ˆ E m j ~X t Ḓ E X 0 j B t m ~X t s e m ; e m ;... where B is the backward shift operator, that is, BX t ˆ X t7. By Kolmogorov's zero-one law, m m ) tends to a constant as m!. It is easy to see that the limit of m m ) is the mean of X t. Similarly, we can show that M m ) tends to the median of X t and that AR m ) tends to the mean of X t. Thus, we proved the following theorem. Theorem 6 In addition to the condition of Theorem, we assume that the time series {X t } is stationary and causal. When m is large, for the conditional mean nonlinear forecast, we have F 0I R Š F 0I R Š RMSE m! and for the conditional median nonlinear forecast, 8 >< F j mean medium j if median < mean! ˆ if median ˆ mean >: F j mean median j if median > mean 8 9 RMAE m! j R mean median x mean df m x j ~X t E j X t median j Note that the F(0) in model (8) may not be / since F is centerized at the mean, whereas F(0) ˆ / in model (9) since F is centerized at the median. In the following, we give an example to illustrate the results in this section. J. forecast, Vol. 6, 49±508 (997) # 997 John Wiley & Sons, Ltd.

Nonlinear AR Models 497 Example Suppose that ~X t ˆ (X,..., X t ) is observed from model () and an AR() model is tted to the data. The autoregressive coe cient f is estimated by the least squares estimate ^f ˆ P t iˆ X i X i P t iˆ X i The linear conditional mean forecast is AR m ˆ ^f m X t and the nonlinear forecast is m m ) ˆ ab m 7 sign ), where b ˆ 7 F(a). The conditional distribution of X t m given ~X t is F m x j ~X t ˆ F x a b m sign F x a b m sign Š. From this, the conditional median forecast M m ) can be evaluated. Figures and give the plots of m m ), M m ) and AR m ) versus X t, with a ˆ.5. The forecast distance of two conditional means is x m ˆ j ^f m X t ab m sign j. Note that the least squares estimate ^f converges to r ˆ E X X E X a j a ab ˆ a as t!, where j is the density function of the standard normal distribution. If a ˆ.5, then ^f 0:7. Figure 3 is the plot of x m ) versus X t, Figures 4±6 the plots of ) (given by Figure. Conditional mean and conditional median forecasts versus origins X for the AR() model and model () with a ˆ.5: odd-step forecast. ÐÐÐ Conditional mean,....... conditional median Figure. Even-step forecast # 997 John Wiley & Sons, Ltd. J. forecast, Vol. 6, 49±508 (997)

498 Meihui Guo and Y. K. Tseng Figure 3 Figure 4 J. forecast, Vol. 6, 49±508 (997) # 997 John Wiley & Sons, Ltd.

Nonlinear AR Models 499 Figure 5 Figure 6 # 997 John Wiley & Sons, Ltd. J. forecast, Vol. 6, 49±508 (997)

500 Meihui Guo and Y. K. Tseng Figure 7. RMSE m ) versus X t for the AR() model and model () with a ˆ.5 Corollary 4) versus X t (the de nition of the modi ed forecast is given in Theorem 7) and Figure 7 the plot of RMSE m ) versus X t, where x RMSE m ˆ m a b m Clearly, from Figure 7 it can be seen that the RMSE m ) is not necessarily a decreasing function in m at all origins X t. This is contrary to Pemberton's observation about RMSE m ) as mentioned earlier. When t is large, the intersection of the two m-step conditional means given X t (by equating the two conditional means) is given by jx t j ˆ abm ^j m a a m b a! D a as m! where D a ˆ aj a 0; b When a ˆ.5, the stationary density f x ˆ j x a j x a J. forecast, Vol. 6, 49±508 (997) # 997 John Wiley & Sons, Ltd.

Nonlinear AR Models 50 Table I. The intersection points of model () with a ˆ.5 and AR() model when X t 4 0 Step m 3 4 5 6 7 8 9 0 ab m 7 / ^f m.08.5 3.0 3.64 4.38 5.8 6.36 7.66 9.3. of model () has modes around x ˆ +.5 and the intersection points for x t 4 0 are given in Table I. Since the intersection points of the two forecasts are not in the neighbourhood of the modes of the stationary density function, contrary to Pemberton's observation (see above), the poor forecast region is not in the region of high density. See Figures 8± for the plots of ) of m m ) versus the stationary density of X t. However, if a is large but m is not, then b m = ^f m and ab m = ^f m a. For example, in Pemberton's experiment a ˆ 3, the intersection points are close to the modes X t ˆ + 3 of the stationary p.d.f. Thus, Pemberton observe a minimum of ) and RMSE m ) in the vicinity of the modes when m is small. Furthermore, since for all X t, x m )! 0 as m!, thus! = and RMSE m )! as m!, which implies the multi-step forecast performance of linear model is asymptotically as good as that of the true nonlinear piecewise constant model. A NEWLY PROPOSED FORECAST Calculating the nonlinear conditional mean forecast, in general, is much easier than computing the conditional median forecast. However, the former is sometimes worse than the AR predictor whose computation is the easiest among the three. If one wants to use the nonlinear forecast for Figure 8 # 997 John Wiley & Sons, Ltd. J. forecast, Vol. 6, 49±508 (997)

50 Meihui Guo and Y. K. Tseng Figure 9 Figure 0 J. forecast, Vol. 6, 49±508 (997) # 997 John Wiley & Sons, Ltd.

Nonlinear AR Models 503 Figure not losing the bene ts of the nonlinear modelling, we suggest a new forecast which has the highest comparison probability relative to the AR forecast. Consider the case when the true model is a nonlinear one. Then the m-step conditional probability distribution function of X t m given ~X t is usually skewed. Thus, the MMSE forecast is liable to be in uenced by the skewness of the conditional distribution; furthermore, its ) is sometimes less than 0.5. For the MMAE forecast, although its comparison probability ) is always greater than or equal to 0.5, yet its value is more di cult to compute. Therefore, we suggest an alternative forecast, combining the MMSE forecast, and the MMAE forecast, which has a higher comparison probability ). The proposed forecast MF m ) is de ned as MF m ˆ m m I R Š M m I R Š if m m > M m ˆ M m I R Š m m I R Š if m m 4 M m where R ˆ { ~X t : m m ) 4 AR m )} and R ˆ { ~X t : m m ) > AR m )}. If the true model is nonlinear, theoretically we can still nd a reasonable approximate linear model. Then, the new forecast is the nonlinear MMSE forecast unless its relative forecast performance is worse than the MMAE forecast, which certainly includes the case when the MMSE performs worse than the linear forecast. In Example, the MF m ) forecast of model () is shown to be the MMSE forecast for ~X t lying in a high stationary density region. In Theorem 7, assuming the true model is nonlinear, ) of the forecast MF m ) is shown to be the maximum of those of MMSE and MMAE forecasts. # 997 John Wiley & Sons, Ltd. J. forecast, Vol. 6, 49±508 (997)

504 Meihui Guo and Y. K. Tseng To indicate the predictor used in the comparison probability, let, m m ),, M m ) and, MF m ) denote the ) of m m ), M m ) and MF m ) respectively. Theorem 7 If model (3) is the true model, then ; MF m ˆ maxf ; m m ; ; M m g 5 Remark It seems that the computation of MF m is more complicated than computing M m ) for M m ) is involved in the de nition of MF m. However, in many cases, the relation between m m ) and M m ) can be determined before computing them (see Example below). Thus, M m ) needs only be computed when MF m ˆ M m. Example (Example, continued) The m-step conditional density of X t m given X t for model () is f m x j X t ˆ k m j x sign a k m j x sign a where b ˆ 7 F(a), k m and k m are given in Corollary 4. The skewness of f m (x j x t ) is a ected by both m and b which determine the weights of the linear combination of j(x a) and j(x 7 a). See Figures and 3 for the plots of f m (x j x t 4 0) versus x when a ˆ,, m ˆ,, 3, 4, 5. It can be proved (see the Appendix) that () m m ) 4 M m ) if X t 4 0 and m is odd or X t > 0 and m is even () m m ) > M m ) if X t 4 0 and m is even or X t > 0 and m is odd. Figure J. forecast, Vol. 6, 49±508 (997) # 997 John Wiley & Sons, Ltd.

Nonlinear AR Models 505 Figure 3 The new forecast of model () with a ˆ.5 can be simpli ed as MF m ˆ m m I Km Š M m I K c m Š, where m m ) ˆ abm 7 sign ), K m ˆ fx t : jx t j 4 jab m = ^f m jg, K c m ˆ fx t : jx t j > jab m = ^f m jg and ^f 0:7. See Table I in Example for the values of ab m = ^f m for m ˆ,,..., 0. Therefore the proposed forecast MF m ) is the m-step conditional mean for origins X t in high stationary probability density regions and is the m-step conditional median for origins X t in low stationary density regions. The plots of ) versus X t for MF m ), m m ) and M m ), m ˆ,, 3 are given in Figures 4±6. DISCUSSION AND CONCLUSIONS If the true model is completely known or partially known (or known functional form with unknown parameters) and the criterion is chosen as the RMSE (RMAE), then the best forecast is the conditional mean (median) forecast. However, if the comparison probability ) is chosen as the criterion, the case becomes complicated. In the theorems in the third section, the expressions of ) are given. It has been proven that ) is an increasing function of the forecast distance x m ) for origin ~X t in disjoint regions. To determine which predictor should be used, one has to compute the two predictors in comparison. This is seemingly not a good deal because the purpose is to use the linear forecast to save computation time. However, there may be alternative ways to determine which predictor should be used. As illustrated in Example, when computing the MF m ), one can determine whether M m ) is needed. Also, as noted by # 997 John Wiley & Sons, Ltd. J. forecast, Vol. 6, 49±508 (997)

506 Meihui Guo and Y. K. Tseng Pemberton, when ~X t is in the region of high density values, the distance x m ( X t ) generally (although not always) has smaller values and hence the linear and nonlinear forecasts are comparable. The newly proposed forecast provides an alternative way of selecting conditional mean or conditional median forecast when the m-step conditional distribution is skewed. However, when the m-step conditional distribution is not of a location scale family or when the model is more complicated than the piecewise constant SETAR model the property of ) still needs further studies. APPENDIX Proof P of Theorem 3 Since the conditional distribution of X t m given X t c j ; c j Š is F m;j x ˆ l iˆ k m ij F x a i (see Pemberton, 990), " # F m;j ˆ Xl k m xm X ij F t m m a i I R Š iˆ " # k m ij F m m a i Xl iˆ x m Given X t c j ; c j Š; m m ˆ Pl iˆ k m ij a i. Therefore, the result is obtained by setting a m ij ˆ m m a i and by model (5). QED Proof of Corollary 4 The conditional distribution of X t m given X t is Therefore, by model (5), for X t in R F m x ˆ k m F x sign a k m F x sign a I R Š ˆ F m ˆ k m F sign a k m F sign a ˆ k m F x m ab m sign sign a k m F x m ab m sign sign a ˆ k m F x m a sign k m k m F x m a sign k m ˆ H m x m A Similarly, one can prove for X t in R Proof of Theorem 5 Since ˆ F m ˆ H m x m A QED E m AR m j ~X t Š ˆ MSE m m j ~X t AR m m m J. forecast, Vol. 6, 49±508 (997) # 997 John Wiley & Sons, Ltd.

Nonlinear AR Models 507 it is easy to derive model (8). Furthermore, since E j x a j E j x b j ˆ b a F a ˆ a b F a Z b a Z a thus model (9) is derived by letting a ˆ M m ) and b ˆ AR m ). Proof of Theorem 7 First consider the case when m m ) > M m ). () If AR m ) 5 m m ), then MF m ) ˆ m m ) and b b x df x; if b > a x b df x; if b < a ; MF m ˆ F m m m j ~X t > F m M m j ~X t 5 ˆ ; M m where F m (. j ~X t ) is the conditional distribution of X t m given ~X t, Q m m m ˆ m m AR m and Q m M m ˆ M m AR m. () If M m ) 4 AR m ) < m m ), then MF m ) ˆ M m ) and ; MF m ˆ F m M m j ~X t 5 F m M m j ~X t ˆ F m M m j ~X t ˆ > F m m m j ~X t ˆ ; m m (3) If AR m ) < M m ), then MF m ) ˆ M m ) and ; MF m ˆ F m M m j ~X t 5 > F m m m j ~X t ˆ ; m m When m m ) 4 M m ), the result can be obtained similarly. QED Proof of () and () in Example Since j(x) is decreasing for x 5 0, Z ak j x dx ˆ k k F ak 0 is decreasing in k. First consider the case for X t 4 0. Let F m (. j X t ) denote the conditional distribution of X t m given X t, then F m m m j X t ˆ k m F ak m k m F ak m " ˆ k m k m F ak m F ak m # k m 5 0 if and only if k m 4 k m k m # 997 John Wiley & Sons, Ltd. J. forecast, Vol. 6, 49±508 (997)

508 Meihui Guo and Y. K. Tseng Thus, m m 5 M m, F m m m j X t 5, k m 4 k m, m is even, since b < 0. The case when X t > 0 can be similarly proved. QED ACKNOWLEDGEMENTS The authors are grateful to Zhidong Bai for his helpful comments and suggestions. The work presented in this paper was supported by the National Science Council of the Republic of China, Grant No. NSC 8-008-N-0-064. REFERENCES Davies, N., Petruccelli, J. D. and Pemberton, J., `An automatic procedure for identi cation, estimation and forecasting self exciting threshold autoregressive models', Statistician, 37 (988), 99±04. Oster, G. and Ipaktchi, A., `Population cycles', Theor. Chem. Periodicity, in Eyring, H. and Henderson, D. (eds), Chem & Biol, New York: Academic Press, ±3. Ozaki, T., `Nonlinear time series models and dynamical systems', in Hannan, E. J., Krishnaiah, P. R. and Rao, M. M. (eds), Handbook of Statistics 5, Amsterdam: North-Holland, 985, 5±83. Pemberton, J., `Forecast accuracy of nonlinear time series models', Tech. Rep. 77, Statistics Research Center, Graduate School of Business, University of Chicago, 989. Pemberton, J., `Piecewise constant autoregressive models for univariate time series', Tech. Rep. MCS-90-04, Department of Mathematics & Computer Science, University of Salford, 990. Priestly, M. B., Spectral Analysis and Time Series, New York: Academic Press, 989. Tiao, G. C. and Tsay, R. S., `Some advances in nonlinear and adaptive modeling in time series', J. Forecast., 3 (994), 09±3. Tong, H., Non-linear Time Series: A dynamical systems approach, Oxford: Oxford University Press, 990. Authors' address: Meihui Guo and Y. K. Tseng, Department of Applied Mathematics, National Sun Yat-sen University, Kaohsiung, ROC 8044. J. forecast, Vol. 6, 49±508 (997) # 997 John Wiley & Sons, Ltd.