Iterative ARIMA-Multiple Support Vector Regression models for long term time series prediction

Similar documents
NON-FIXED AND ASYMMETRICAL MARGIN APPROACH TO STOCK MARKET PREDICTION USING SUPPORT VECTOR REGRESSION. Haiqin Yang, Irwin King and Laiwan Chan

Support Vector Machine Regression for Volatile Stock Market Prediction

Do we need Experts for Time Series Forecasting?

Improved the Forecasting of ANN-ARIMA Model Performance: A Case Study of Water Quality at the Offshore Kuala Terengganu, Terengganu, Malaysia

Outliers Treatment in Support Vector Regression for Financial Time Series Prediction

Estimation and application of best ARIMA model for forecasting the uranium price.

Neural network time series classification of changes in nuclear power plant processes

Time Series Analysis -- An Introduction -- AMS 586

Based on SARIMA-BP hybrid model and SSVM model of international crude oil price prediction research

A Wavelet Neural Network Forecasting Model Based On ARIMA

A Machine Learning Approach to Define Weights for Linear Combination of Forecasts

On the benefit of using time series features for choosing a forecasting method

UNIVARIATE TIME SERIES ANALYSIS BRIEFING 1970

Foreign Exchange Rates Forecasting with a C-Ascending Least Squares Support Vector Regression Model

TIME SERIES ANALYSIS AND FORECASTING USING THE STATISTICAL MODEL ARIMA

Journal of of Computer Applications Research Research and Development and Development (JCARD), ISSN (Print), ISSN

FORECASTING THE INVENTORY LEVEL OF MAGNETIC CARDS IN TOLLING SYSTEM

Nonlinear System Identification using Support Vector Regression

Forecast daily indices of solar activity, F10.7, using support vector regression method

Sensitivity Analysis for Time Lag Selection to Forecast Seasonal Time Series using Neural Networks and Support Vector Machines

Space-Time Kernels. Dr. Jiaqiu Wang, Dr. Tao Cheng James Haworth University College London

EM-algorithm for Training of State-space Models with Application to Time Series Prediction

SVRG++ with Non-uniform Sampling

Probabilistic Regression Using Basis Function Models

TIME SERIES DATA PREDICTION OF NATURAL GAS CONSUMPTION USING ARIMA MODEL

Prashant Pant 1, Achal Garg 2 1,2 Engineer, Keppel Offshore and Marine Engineering India Pvt. Ltd, Mumbai. IJRASET 2013: All Rights are Reserved 356

ADDING EMD PROCESS AND FILTERING ANALYSIS TO ENHANCE PERFORMANCES OF ARIMA MODEL WHEN TIME SERIES IS MEASUREMENT DATA

Temperature Forecast in Buildings Using Machine Learning Techniques

FORECASTING SAVING DEPOSIT IN MALAYSIAN ISLAMIC BANKING: COMPARISON BETWEEN ARTIFICIAL NEURAL NETWORK AND ARIMA

Design of Time Series Model for Road Accident Fatal Death in Tamilnadu

Support Vector Machine for Classification and Regression

Support Vector Ordinal Regression using Privileged Information

Load Forecasting Using Artificial Neural Networks and Support Vector Regression

SARIMA-ELM Hybrid Model for Forecasting Tourist in Nepal

Selection of the Appropriate Lag Structure of Foreign Exchange Rates Forecasting Based on Autocorrelation Coefficient

DEPARTMENT OF ECONOMETRICS AND BUSINESS STATISTICS

Time Series Forecasting: A Tool for Out - Sample Model Selection and Evaluation

Stochastic Gradient Estimate Variance in Contrastive Divergence and Persistent Contrastive Divergence

Support Vector Regression (SVR) Descriptions of SVR in this discussion follow that in Refs. (2, 6, 7, 8, 9). The literature

7. Forecasting with ARIMA models

{ } Stochastic processes. Models for time series. Specification of a process. Specification of a process. , X t3. ,...X tn }

Lesson 13: Box-Jenkins Modeling Strategy for building ARMA models

Comparison Forecasting with Double Exponential Smoothing and Artificial Neural Network to Predict the Price of Sugar

A Support Vector Regression Model for Forecasting Rainfall

Package TSPred. April 5, 2017

A Comparison of the Forecast Performance of. Double Seasonal ARIMA and Double Seasonal. ARFIMA Models of Electricity Load Demand

Predicting Time of Peak Foreign Exchange Rates. Charles Mulemi, Lucio Dery 0. ABSTRACT

Research Article Stacked Heterogeneous Neural Networks for Time Series Forecasting

Resampling techniques for statistical modeling

arxiv: v1 [stat.co] 11 Dec 2012

Analysis of Fast Input Selection: Application in Time Series Prediction

Chapter 12: An introduction to Time Series Analysis. Chapter 12: An introduction to Time Series Analysis

A CUSUM approach for online change-point detection on curve sequences

arxiv: v1 [cs.lg] 21 Jul 2016

FORECASTING SUGARCANE PRODUCTION IN INDIA WITH ARIMA MODEL

FORECASTING OF ECONOMIC QUANTITIES USING FUZZY AUTOREGRESSIVE MODEL AND FUZZY NEURAL NETWORK

Solar irradiance forecasting for Chulalongkorn University location using time series models

Probabilistic Energy Forecasting

Deep Learning Architecture for Univariate Time Series Forecasting

Immediate Reward Reinforcement Learning for Projective Kernel Methods

Exercises - Time series analysis

RAINFALL RUNOFF MODELING USING SUPPORT VECTOR REGRESSION AND ARTIFICIAL NEURAL NETWORKS

Long-Term Prediction of Time Series by combining Direct and MIMO Strategies

Linear Dependency Between and the Input Noise in -Support Vector Regression

Econometric Forecasting

Modelling Wind Farm Data and the Short Term Prediction of Wind Speeds

Linear and Non-Linear Dimensionality Reduction

Journal of Asian Scientific Research PREDICTION OF FATAL ROAD TRAFFIC CRASHES IN IRAN USING THE BOX-JENKINS TIME SERIES MODEL. Ayad Bahadori Monfared

A Revisit to Support Vector Data Description (SVDD)

Forecasting Stock Prices using Hidden Markov Models and Support Vector Regression with Firefly Algorithm

Economic Modelling 29 (2012) Contents lists available at SciVerse ScienceDirect. Economic Modelling

Research Article Forecasting Air Passenger Traffic by Support Vector Machines with Ensemble Empirical Mode Decomposition and Slope-Based Method

Soft Sensor Modelling based on Just-in-Time Learning and Bagging-PLS for Fermentation Processes

Forecasting USD/IQD Future Values According to Minimum RMSE Rate

Square Penalty Support Vector Regression

Nonlinear Characterization of Activity Dynamics in Online Collaboration Websites

Prediction of Seasonal Rainfall Data in India using Fuzzy Stochastic Modelling

An Improved Conjugate Gradient Scheme to the Solution of Least Squares SVM

Dynamic Time Series Regression: A Panacea for Spurious Correlations

Discussion of Some Problems About Nonlinear Time Series Prediction Using ν-support Vector Machine

Time Series Outlier Detection

Nonparametric time series forecasting with dynamic updating

FORECASTING. Methods and Applications. Third Edition. Spyros Makridakis. European Institute of Business Administration (INSEAD) Steven C Wheelwright

Iteratively Reweighted Least Square for Asymmetric L 2 -Loss Support Vector Regression

Forecasting Network Activities Using ARIMA Method

Time Series and Forecasting

Forecasting: Methods and Applications

Electrical Energy Modeling In Y2E2 Building Based On Distributed Sensors Information

Predicting the Probability of Correct Classification

An Anomaly Detection Method for Spacecraft using Relevance Vector Learning

Some Time-Series Models

Minitab Project Report - Assignment 6

Using Analysis of Time Series to Forecast numbers of The Patients with Malignant Tumors in Anbar Provinc

FORECASTING YIELD PER HECTARE OF RICE IN ANDHRA PRADESH

Forecasting strong seasonal time series with artificial neural networks

A hybrid model of stock price prediction based on the PCA-ARIMA-BP

FE570 Financial Markets and Trading. Stevens Institute of Technology

An introduction to Support Vector Machines

(SARIMA) SARIMA 1390 **

Is the Basis of the Stock Index Futures Markets Nonlinear?

Transcription:

and Machine Learning Bruges (Belgium), 23-25 April 24, i6doccom publ, ISBN 978-2874995-7 Available from http://wwwi6doccom/fr/livre/?gcoi=2843244 Iterative ARIMA-Multiple Support Vector Regression models for long term time series prediction João Fausto Lorenzato de Oliveira and Teresa B Ludermir Federal University of Pernambuco - Center of informatics Av Jornalista Anibal Fernandes, s/n, Recife, PE, 574-56, Brazil Abstract Support Vector Regression (SVR) has been widely applied in time series forecasting Considering long term predictions, iterative predictions perform many one-step-ahead predictions until the desired horizon is achieved This process accumulates the error from previous predictions and may affect the quality of forecasts In order to improve long term iterative predictions a hybrid multiple Autoregressive Integrated Moving Average(ARIMA)-SVR model is applied to perform predictions considering linear and non-linear components from the time series The results show that the proposed method produces more accurate predictions in the long term context Introduction A time series is a discrete-time stochastic process composed by a finite set of items correlated in time Time series forecasting has many applications in areas such as planning, management, production, maintenance and control of industrial processes The forecasting process is based on some knowledge of the past, which can be performed by several statistical methods [] and by nonlinear methods such as Support Vector Regression (SVR) [2] and Artificial Neural Networks (ANNs) [3] Some time series may present seasonal or\and trend patterns, which must be modeled by a method in order to produce better forecasts Statistical methods such as ARIMA are able to map linear aspects such as trend, however nonlinear patterns are not easily captured Different models can complement each other in capturing linear or nonlinear patterns of data sets in time series Bates and Granger [4] explored several combinations of architectures for time series prediction, Pai and Lin used a hybrid ARIMA-SVR model to predict stock prices [5], Zhu and Wei used a hybrid ARIMA-Least Squares Support Vector Regression(LSSVR) to predict carbon price However, when performing iterative multi-step-ahead predictions, the accumulation of error may still be present at each forecast iteration In order to address this problem, Zhang et al used a multiple SVR [6] strategy where the SVR models are trained independently in different prediction horizons This method reduces the number of iterations and may reduce the propagated error by performing many direct predictions The authors would like to thank FACEPE, CNPq and CAPES (Brazilian Research Agencies) for their financial support 27

and Machine Learning Bruges (Belgium), 23-25 April 24, i6doccom publ, ISBN 978-2874995-7 Available from http://wwwi6doccom/fr/livre/?gcoi=2843244 In this paper we propose a Multiple ARIMA-SVR model in order to map linear and nonlinear patterns in time series and to improve the accuracy in iterative long term forecasts The remainder of this paper is organized as follows: Section 2 presents a brief review of time series forecasting strategies, Section 3 shows the proposed method Section 4 presents the experiments and the conclusions and future works are presented in section 5 2 Time series forecasting In this section, the iterated, direct and multiple approaches for time series forecasting are reviewed For all three approaches consider a time series {z t } N t=, where N is the length of the time series In the one-step-ahead prediction the model depends on the d past values and has the form z (t + ) = f(z(t), z(t ),, z(t (d ))), where d is the embedding dimension or time window The prediction of p-steps-ahead in the iterative approach takes the form z (t + h) = f(z (t + h ), z (t + h 2),, z (t + h d)) [7], where the past d predictions are taken into consideration in order to perform the p th prediction In the iterative approach, the model is trained to perform one-step-ahead predictions at time t + iteratively h times using past predictions as inputs for the next prediction This process do not need information at the desired horizon h however it accumulates the error from all the past predictions Only one model is required to perform predictions The prediction of p-steps-ahead in the direct approach has the form z (t + h) = f(z(t), z(t ),, z(t (d ))) [8] In other words, the model is trained directly at the horizon h, and the predictions are performed directly into that horizon In this approach we need information at horizon h in order to train the model and the error is not accumulated To perform predictions in different values of h, other models have to be trained at this horizon The multiple approach takes into consideration the iterative and direct methods [6] Multiple models are trained independently at different horizons, and perform iterative predictions until the desired horizon is achieved Given a horizon h, k models (with k < h) will be trained independently to perform predictions at the horizons {t +, t + 2,, t + k} This method does not need to train h models as needed in the direct approach and produces less accumulation of error than the iterative method 3 Proposed Method The proposed method incorporates the linear and nonlinear mappings of the time series through a composition of a ARIMA model and multiple SVR models The ARIMA model was first introduced by Box and Jenkins [9] and has been widely applied in time series forecasting The predictions in the ARIMA model are a linear combination of past values and past residual errors, and has the form presented on eq 28

and Machine Learning Bruges (Belgium), 23-25 April 24, i6doccom publ, ISBN 978-2874995-7 Available from http://wwwi6doccom/fr/livre/?gcoi=2843244 θ p (B)( B) s z t = ϕ q (B)ϵ t () where z t is the actual value and ϵ t is the residual error at time t, B is a backward shift operator (Bz t = z t and z t = z t z t = ( B)z t ), ϕ i and θ j are the coefficients, p and q are degrees for the autoregressive and moving average polynomial functions The integration part is due to the differencing performed to the time series in order to obtain a stationary series The number of differences is determined by the parameter s The SVR method was proposed by Vapnik [] and it is based on the structured risk minimization (SRM) principle, aiming at minimizing an upper bound of the generalization error Let {z i, y i } l i= be a training set where z Rd, y i R is the prediction output of z i, d is the embbeding dimension of the time series and l is the number of training samples The objective of SVR is to find the best function from a set of possible functions {f f(z) = w T z + b, w R d, b R}, where w is the weight vector and b is a bias or threshold In order to find the best function f, it is necessary to minimize the regularized risk 2 w 2 + C l i= L(y i, f(z i )), where C > is a regularization factor, is a 2-norm and L(, ) is a loss function In order to induce sparsity in SVR the ϵ-insensitive loss function presented on eq 2, which creates an ϵ-tube allowing some predictions to fall within the boundaries of this ϵ-tube {, f(z) y < ϵ L(y, f(z)) = (2) f(z) y ϵ, otherwise The nonlinearity of SVRs is achieved by mapping the input space into a higher dimensionality space This mapping is possible through the use of kernels, and the regression function takes the form f(z) = l i= (α i αi )k(z i, z) + b, where α and α are lagrange multipliers and k(z i, z) is a kernel function In this study, the gaussian kernel is adopted (k(z i, z j ) = exp( z i z j 2 2γ )), 2 where γ is a parameter of the gaussian kernel The proposed model uses the ARIMA method as a prepossessing step in order to map linear patterns in the time series z(t) z(t-) z(t-d+) ARIMA e(t) e(t-) e(t-d+) SVR SVR 2 SVR k e (t+) e (t+2) e (t+k) SVR SVR 2 SVR k SVR SVR 2 SVR k e (t+(r-)k+) z (t+2) e (t+(r-)k+2) e (t+(r-)k+k) Fig : Hybrid Arima Multiple SVR Method Fig shows the framework of the ARIMA Multiple SVR method (ASVR-M) The r variable denotes the number of iterations of the method The forecasting 29

and Machine Learning Bruges (Belgium), 23-25 April 24, i6doccom publ, ISBN 978-2874995-7 Available from http://wwwi6doccom/fr/livre/?gcoi=2843244 procedure of this method performs less iterations than the iterative method, which runs as many iterations as the horizon size The residuals produced by the ARIMA e t = z t z t, are used as an input for the Multiple SVR methods In this step, k SVR models are trained independently, in order to perform iterative predictions The final forecast can be achieved by adding the forecasts of the ARIMA and multiple SVR models o t = z t + e t 4 Experiments The experiments were performed in four benchmark datasets including Sunspot, Airline Passengers, IBM stock prices and Death Our proposed approach is compared with other methods such as Iterative SVR (SVR-I), Direct SVR (SVR-D), Multiple SVR (SVR-M), ARIMA Iterative SVR (ASVR-I) and ARIMA Direct SVR (ASVR-D), implemented using the LIBSVM package [] in a MATLAB 82 environment The datasets were partitioned, using 8% for training and 2% for testing First the p, q and s parameters of the ARIMA(p,s,q) model are defined with grid search 5-fold cross-validation in the training data with parameters ranging from {, 2}, {,, 2} and {,, 2} respectively and the best ARIMA(p,s,q) model is selected based on the Aikake s Information Criterion (AIC) [2] In all SVR models the parameter selection was performed by doing a -fold cross-validation grid search in the residuals produced by the selected ARIMA model with C, ϵ and γ within the range {,,, 2 }, { 4, 3, 2 } and {2 5, 2 4,, 2 2 } respectively on the training set according to [6] The annual Wolf s sunspot numbers dataset has 289 points and presents seasonal patterns The monthly airline passengers dataset has 44 points and presents trend and seasonal patterns The death dataset represents the number of deaths and serious injuries in UK road accidents each month and has 92 points The daily IBM stock closing prices has 369 points and presents trend patterns Fig 2 presents the autocorrelograms for all datasets, where the lags are set to unity The sunspot, passengers, IBM and death datasets are available from [3] The experiments for the sunspot, airlines passenger, death and IBM were performed considering an embedding dimension d = 6, prediction horizon h = and number of SVR models in the multiple approaches k = d based on [6] All datasets had their values scaled to [] The experiments were executed for iterations The performance measure used in this work as the mean squared error (MSE) described on eq 3 N i= MSE = (z i z i) 2 (3) N Table contemplates the mean and standard deviations for the MSEs in all datasets The results show that in most cases the hybrid ARIMA-SVR procedure 2

and Machine Learning Bruges (Belgium), 23-25 April 24, i6doccom publ, ISBN 978-2874995-7 Available from http://wwwi6doccom/fr/livre/?gcoi=2843244 Sunspot Passenger 5 5 5 2 4 6 Death 5 2 4 6 IBM 5 5 5 2 4 6 5 5 Fig 2: Autocorrelograms of Datasets Method Sunspot Passenger Death IBM SVR-I 64(3) 36(28) 9(3) 4() SVR-D 52(5) 2(4) 23() 8(4) SVR-M 6(3) 53(3) 64(6) (3) ASVR-I 3(4) 2(26) 25(3) 4() ASVR-D 32(2) 3() 25(3) 7() ASVR-M 45(5) 26(3) 6(6) 3() Table : Mean square errors and Standard deviations for all datasets improved the values achieved by the SVR strategies for long-term time series forecasting An empirical analysis of the results achieved demonstrates that in the sunspot dataset the ASVR-I achieved the best results, followed by the ASVR-D In the Passenger dataset, the ASVR-D outperformed the other hybrid approaches, however the ASVR-I presented a high variance in the results, while the results obtained by the ASVR-M were more stable In the IBM dataset the ASVR-M achieve the best results, followed by the ASVR-I method In the Death dataset the ASVR-D and SVR-D achieved the lowest errors, followed by the ASVR-M method A Wilcoxon signed-rank test was applied to perform a robust analysis with 95% of confidence In the Sunspot dataset, the ASVR-I and ASVR-D achieved similar results and performed best In the Passenger dataset the ASVR-D outperformed all techniques and the ASVR-I and ASVR-M performed similarly In the Death dataset the ASVR-D and SVR-D achieved similar performance In 2

and Machine Learning Bruges (Belgium), 23-25 April 24, i6doccom publ, ISBN 978-2874995-7 Available from http://wwwi6doccom/fr/livre/?gcoi=2843244 the IBM dataset the ASVR-M and SVR-I achieved similar results 5 Conclusion This paper realized a comparison between strategies (iterative, direct, multiple) to perform long-term time series forecasting using Support Vector Regression (SVR) models and proposed a hybrid model based on a Autoregressive Integrated Moving Average (ARIMA) model and on multiple SVRs The direct approaches (SVR-D, ASVR-D) achieved the best results in most datasets, which is expected since there is no iterative error accumulation However the multiple approaches (SVR-M and ASVR-M) tends to reduce the accumulation error by running fewer iterations Under these settings the hybridization increased the performance of most methods, however the iterative method was the one with more increase of performance, in most cases achieving similar results to the ASVR-M For future works, we intend to analyze the performance of a hybridization between ARIMA and an ensemble of SVRs to perform prediction using the iterative, direct and multiple strategies References [] Spyros Makridakis, Steven C Wheelwright, and Rob J Hyndman Forecasting methods and applications Wiley com, 28 [2] K-R Müller, Alex J Smola, Gunnar Rätsch, Bernhard Schölkopf, Jens Kohlmorgen, and Vladimir Vapnik Predicting time series with support vector machines In Artificial Neural Networks ICANN 97, pages 999 4 Springer, 997 [3] Guoqiang Zhang, B Eddy Patuwo, and Michael Y Hu Forecasting with artificial neural networks:: The state of the art International journal of forecasting, 4():35 62, 998 [4] John M Bates and Clive WJ Granger The combination of forecasts OR, pages 45 468, 969 [5] Ping-Feng Pai and Chih-Sheng Lin A hybrid arima and support vector machines model in stock price forecasting Omega, 33(6):497 55, 25 [6] Li Zhang, Wei-Da Zhou, Pei-Chann Chang, Ji-Wen Yang, and Fan-Zhang Li Iterated time series prediction with multiple support vector regression models Neurocomputing, 22 [7] Abdelhamid Bouchachia and Saliha Bouchachia Ensemble learning for time series prediction In Proceedings of the International Workshop on Nonlinear Dynamic Systems and Synchronization, 28 [8] Souhaib Ben Taieb, Antti Sorjamaa, and Gianluca Bontempi Multiple-output modeling for multi-step-ahead time series forecasting Neurocomputing, 73():95 957, 2 [9] George EP Box, Gwilym M Jenkins, and Gregory C Reinsel Time series analysis: forecasting and control Wiley com, 23 [] Vladimir Vapnik The nature of statistical learning theory springer, 2 [] Chih-Chung Chang and Chih-Jen Lin LIBSVM: A library for support vector machines ACM Transactions on Intelligent Systems and Technology, 2:27: 27:27, 2 [2] Hamparsum Bozdogan Model selection and akaike s information criterion (aic): The general theory and its analytical extensions Psychometrika, 52(3):345 37, 987 [3] RJ Hyndman Time series data library, 2 22