Pattern Matching and Neural Networks based Hybrid Forecasting System

Similar documents
A LONG MEMORY PATTERN MODELLING AND RECOGNITION SYSTEM FOR FINANCIAL TIME-SERIES FORECASTING

Electric Load Forecasting Using Wavelet Transform and Extreme Learning Machine

A FUZZY NEURAL NETWORK MODEL FOR FORECASTING STOCK PRICE

Analysis of Interest Rate Curves Clustering Using Self-Organising Maps

Artificial Neural Networks Examination, March 2004

FORECASTING OF INFLATION IN BANGLADESH USING ANN MODEL

CHAPTER 6 CONCLUSION AND FUTURE SCOPE

Evolutionary Functional Link Interval Type-2 Fuzzy Neural System for Exchange Rate Prediction

Deep Learning Architecture for Univariate Time Series Forecasting

Direct Method for Training Feed-forward Neural Networks using Batch Extended Kalman Filter for Multi- Step-Ahead Predictions

A Hybrid Model of Wavelet and Neural Network for Short Term Load Forecasting

A Machine Learning Approach to Define Weights for Linear Combination of Forecasts

ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD

Forecasting demand in the National Electricity Market. October 2017

Introduction to Natural Computation. Lecture 9. Multilayer Perceptrons and Backpropagation. Peter Lewis

Efficient Forecasting of Exchange rates with Recurrent FLANN

Forecasting exchange rate volatility using conditional variance models selected by information criteria

epochs epochs

ECLT 5810 Classification Neural Networks. Reference: Data Mining: Concepts and Techniques By J. Hand, M. Kamber, and J. Pei, Morgan Kaufmann

Weight Initialization Methods for Multilayer Feedforward. 1

Forecasting Currency Exchange Rates: Neural Networks and the Random Walk Model

Optimum Neural Network Architecture for Precipitation Prediction of Myanmar

22/04/2014. Economic Research

Forecasting Regional Employment in Germany: A Review of Neural Network Approaches. Objectives:

Forecasting Exchange Rate Change between USD and JPY by Using Dynamic Adaptive Neuron-Fuzzy Logic System

Discussion Papers in Economics. Ali Choudhary (University of Surrey and State Bank of Pakistan) & Adnan Haider (State Bank of Pakistan) DP 08/08

Artificial Neural Networks Examination, June 2005

On methods to assess the significance of community structure in networks of financial time series.

Artificial Neural Networks Examination, June 2004

Non-linear Analysis of Shocks when Financial Markets are Subject to Changes in Regime

ADAPTIVE NEURO-FUZZY INFERENCE SYSTEMS

CSE 352 (AI) LECTURE NOTES Professor Anita Wasilewska. NEURAL NETWORKS Learning

Short Term Load Forecasting Based Artificial Neural Network

Radial Basis Functions Networks to hybrid neuro-genetic RBFΝs in Financial Evaluation of Corporations

MODELLING ENERGY DEMAND FORECASTING USING NEURAL NETWORKS WITH UNIVARIATE TIME SERIES

Integrated Electricity Demand and Price Forecasting

Supervised locally linear embedding

Artificial Neural Networks Examination, March 2002

is used in the empirical trials and then discusses the results. In the nal section we draw together the main conclusions of this study and suggest fut

Negatively Correlated Echo State Networks

Introduction to Neural Networks

Hybrid HMM/MLP models for time series prediction

Distribution Forecasting of Nonlinear Time Series with Associative Memories. A. R. Pasley

Predicting Time of Peak Foreign Exchange Rates. Charles Mulemi, Lucio Dery 0. ABSTRACT

Forecasting River Flow in the USA: A Comparison between Auto-Regression and Neural Network Non-Parametric Models

Multilayer Perceptrons (MLPs)

Introduction to Neural Networks

Machine Learning and Data Mining. Multi-layer Perceptrons & Neural Networks: Basics. Prof. Alexander Ihler

Do we need Experts for Time Series Forecasting?

Electricity Demand Forecasting using Multi-Task Learning

Longshore current velocities prediction: using a neural networks approach

Short-term wind forecasting using artificial neural networks (ANNs)

Numerical Learning Algorithms

Forecasting Crude Oil Price Using Neural Networks

A STATE-SPACE NEURAL NETWORK FOR MODELING DYNAMICAL NONLINEAR SYSTEMS

Module 12. Machine Learning. Version 2 CSE IIT, Kharagpur

Selection of the Appropriate Lag Structure of Foreign Exchange Rates Forecasting Based on Autocorrelation Coefficient

CHAOS THEORY AND EXCHANGE RATE PROBLEM

A Hybrid Method of CART and Artificial Neural Network for Short-term term Load Forecasting in Power Systems

Forecasting Stock Prices using Hidden Markov Models and Support Vector Regression with Firefly Algorithm

Lecture 5: Logistic Regression. Neural Networks

Artificial Intelligence (AI) Common AI Methods. Training. Signals to Perceptrons. Artificial Neural Networks (ANN) Artificial Intelligence

Correlation Network Evolution Using Mean Reversion Autoregression

Ensembles of Nearest Neighbor Forecasts

CHAPTER 4 MODEL TESTING ON STREAMFLOWS: I. ANNUAL MEAN FLOWS

Decision Models Lecture 5 1. Lecture 5. Foreign-Currency Trading Integer Programming Plant-location example Summary and Preparation for next class

A FUZZY TIME SERIES-MARKOV CHAIN MODEL WITH AN APPLICATION TO FORECAST THE EXCHANGE RATE BETWEEN THE TAIWAN AND US DOLLAR.

Advanced statistical methods for data analysis Lecture 2

Introduction to Artificial Neural Networks

A new method for short-term load forecasting based on chaotic time series and neural network

SARIMA-ELM Hybrid Model for Forecasting Tourist in Nepal

Article. Chaotic analysis of predictability versus knowledge discovery techniques: case study of the Polish stock market

An Analysis of the INFFC Cotton Futures Time Series: Lower Bounds and Testbed Design Recommendations

Lecture 7 Artificial neural networks: Supervised learning

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18

A neuro-fuzzy system for portfolio evaluation

Cascade Neural Networks with Node-Decoupled Extended Kalman Filtering

Prediction of Hourly Solar Radiation in Amman-Jordan by Using Artificial Neural Networks

ANN and Statistical Theory Based Forecasting and Analysis of Power System Variables

Keywords- Source coding, Huffman encoding, Artificial neural network, Multilayer perceptron, Backpropagation algorithm

A Wavelet Neural Network Forecasting Model Based On ARIMA

A Novel Activity Detection Method

Application of Artificial Neural Networks in Evaluation and Identification of Electrical Loss in Transformers According to the Energy Consumption

A SEASONAL FUZZY TIME SERIES FORECASTING METHOD BASED ON GUSTAFSON-KESSEL FUZZY CLUSTERING *

AI Programming CS F-20 Neural Networks

ARTIFICIAL INTELLIGENCE. Artificial Neural Networks

COMPARING PERFORMANCE OF NEURAL NETWORKS RECOGNIZING MACHINE GENERATED CHARACTERS

Another Error Measure for Selection of the Best Forecasting Method: The Unbiased Absolute Percentage Error

A New Look at Nonlinear Time Series Prediction with NARX Recurrent Neural Network. José Maria P. Menezes Jr. and Guilherme A.

What Do Neural Networks Do? MLP Lecture 3 Multi-layer networks 1

FORECASTING SAVING DEPOSIT IN MALAYSIAN ISLAMIC BANKING: COMPARISON BETWEEN ARTIFICIAL NEURAL NETWORK AND ARIMA

Artifical Neural Networks

Classifier Selection. Nicholas Ver Hoeve Craig Martek Ben Gardner

Deep Feedforward Networks

Part 8: Neural Networks

Application of Artificial Neural Network for Short Term Load Forecasting

Defining Feedforward Network Architecture. net = newff([pn],[s1 S2... SN],{TF1 TF2... TFN},BTF,LF,PF);

Speaker Representation and Verification Part II. by Vasileios Vasilakakis

Data Mining Part 5. Prediction

Chapter 3 Supervised learning:

Transcription:

Pattern Matching and Neural Networks based Hybrid Forecasting System Sameer Singh and Jonathan Fieldsend PA Research, Department of Computer Science, University of Exeter, Exeter, UK Abstract In this paper we propose a Neural Net- hybrid for forecasting time-series data. The neural network model uses the traditional MLP architecture and backpropagation method of training. Rather than using the last n lags for prediction, the input to the network is determined by the output of the (Pattern Modelling and Recognition System). matches current patterns in the time-series with historic data and generates input for the neural network that consists of both current and historic information. The results of the hybrid model are compared with those of neural networks and on their own. In general, there is no outright winner on all performance measures, however, the hybrid model is a better choice for certain types of data, or on certain error measures. 1. Hybrid Forecasting System The science of forecasting relies heavily on the models used for forecasting, quantity and quality of data, and the ability to pick the right models for a given data. A number of past publications on pattern matching have argued the need for using historic information through pattern matching for forecasting. A number of these pattern matching techniques, such as our previous model, have been very successful on a range of economic and financial data. The main philosophy behind these pattern matching techniques is to identify the best matching past trends to current ones and use the knowledge of how the time series behaved in the past in those situations to make predictions for the future. A range of nearest neighbour strategies can be adopted for this matching such as the fuzzy Single Nearest Neighbour (S) approach. Neural networks do not normally use historic matches for forecasting. Instead, their inputs are taken as recent lags [2]. However, their ability to formulate a non-linear relationship between inputs and output has a considerable advantage for producing accurate forecasts [3].

In this paper we combine the pattern matching ability with the determination of nonlinear relationship between inputs and output by generating a hybrid model for forecasting. The hybrid model uses the model [6,7] for determining historic matches used as inputs to a neural network model. The results are shown on a range of scientific time series data and financial market data. We first describe our and neural network models on a stand-alone basis, and then their combination. 1.1 component If we choose to represent a time-series as y = {y 1, y 2,... y n }, then the current state of size one of the time-series is represented by its current value y n. One simple method of prediction can be based on identifying the closest neighbour of y n in the past data, say y j, and predicting y n+1 on the basis of y j+1. Calculating an average prediction based on more than one nearest neighbour can modify this approach. The definition of the current state of a time-series can be extended to include more than one value, e.g. the current state s c of size two may be defined as {y n-1, y n }. For such a current state, the prediction will depend on the past state s p {y j-1, y j } and next series value y + p given by y j+1, provided that we establish that the state {y j-1, y j } is the nearest neighbour of the state {y n-1, y n } using some similarity measurement. In this paper, we also refer to states as patterns. In theory, we can have a current state of any size but in practice only matching current states of optimal size to past states of the same size yields accurate forecasts since too small or too large neighbourhoods do not generalise well. The optimal state size must be determined experimentally on the basis of achieving minimal errors on standard measures through an iterative procedure. We can formalise the prediction procedure as follows: ÿ = φ(s c, s p, y + p, k, c) where ÿ is the prediction for the next time step, s c is the current state, s p is the nearest past state, y + p is the series value following past state s p, k is the state size and c is the matching constraint. Here ÿ is a real value, s c or s p can be represented as a set of real values, k is a constant representing the number of values in each state, i.e. size of the set, and c is a constraint which is user defined for the matching process. We define c as the condition of matching operation that series direction change for each member in s c and s p is the same. In order to illustrate the matching process for series prediction further, consider the time series as a vector y = {y 1, y 2,... y n } where n is the total number of points in the series. Often, we also represent such a series as a function of time, e.g. y n = y t, y n-1 = y t-1, and so on. A segment in the series is defined as a difference vector δ = (δ 1, δ 2,... δ n-1 ) where δ i = y i+1 - y i, i, 1 i n-1. A pattern contains one or more segments and it can be visualised as a string of segments ρ = (δ i, δ i+1,... δ h ) for given values of i and h, 1 i,h n-1, provided that h>i. In order to define any pattern mathematically, we choose to tag the time series y with a vector of change in direction. For this purpose,

a value y i is tagged with a if y i+1 < y i, and as a 1 if y i+1 y i. Formally, a pattern in the time-series is represented as ρ = (b i, b i+1,... b h ) where b is a binary value. The complete time-series is tagged as (b 1,...b n-1 ). For a total of k segments in a pattern, it is tagged with a string of k b values. For a pattern of size k, the total number of binary patterns (shapes) possible is 2 k. The technique of matching structural primitives is based on the premise that the past repeats itself. It has been noted in previous studies that the dynamic behaviour of time-series can be efficiently predicted by using local approximation. For this purpose, a map between current states and the nearest neighbour past states can be generated for forecasting. Pattern matching in the context of time-series forecasting refers to the process of matching current state of the time series with its past states. Consider the tagged time series (b 1, b i,... b n-1 ). Suppose that we are at time n (y n ) trying to predict y n+1. A pattern of size k is first formulated from the last k tag values in the series, ρ = (b n-k,... b n-1 ). The size k of the structural primitive (pattern) used for matching has a direct effect on the prediction accuracy. Thus the pattern size k must be optimised for obtaining the best results. For this k is increased in every trial by one unit till it reaches a predefined maximum allowed for the experiment and the error measures are noted; the value of k that gives the least error is finally selected. The aim of a pattern matching algorithm is to find the closest match of ρ in the historical data (estimation period) and use this for predicting y n+1. The magnitude and direction of prediction depend on the match found. The success in correctly predicting series depends directly on the pattern matching algorithm. The first step is to select a state/pattern of minimal size (k=2). A nearest neighbour of this pattern is determined from historical data on the basis of smallest offset. There are two cases for prediction: either we predict high or we predict low. The prediction ÿ n+1 is scaled on the basis of the similarity of the match found. We use a number of widely applied error measures for estimating the accuracy of the forecast and selecting optimal k size for minimal error. The forecasting process is repeated with a given test data for states/patterns of size greater than two and a model with smallest k giving minimal error is selected. In our experiments k is iterated between 2 k 5. 1.2 Neural Network component In this paper we use the standard MLP architecture with backpropagation mode of learning. In order to enable consistency between the Neural Network model and the model, the Neural Network inputs are the 6 most recent lags of the time series (i.e. Y t-1, Y t-2, Y t-3, Y t-4, Y t-5 & Y t-6, when the value to be predicted is the actual Y t ). This mimics the approach of pattern matching of up to 5 historic differences (δ values) which therefore uses the information contained in the most recent 6 lags. (In other circumstances the 6 lags chosen would be those with the highest partial autocorrelation function value when correlated with the actual [5]). In our study, neural networks have two hidden layers with 5 sigmoidal nodes in each and the networks are fully connected.

The learning rate was set at.5 with a momentum of.5. The Neural Networks training was stopped when the combined RMSE on the test and training set had fallen by less than.25% of their value 5 epochs ago. The inclusion of the validation error prevents over- fitting, however by summing the two errors (as opposed to strictly using the test error on its own), a slight trade-off is permitted between the errors while pushing through local minima. In addition this stopping operator was only used after at least 1 epochs had passed during training. 1.3 Neural Net- Neural Networks with generated inputs have the same stopping regime as the standard neural network described above with only topological difference in the form of number of inputs. Two hybrid models were trained. The first was training on the matched deltas δ found by the algorithm, the predicted delta and the lags used to find them, i.e. a model fitting two historic deltas would use 3 lags. A single asterisk in Table 2 denotes this model (*). The second model used the matched deltas δ found by the algorithm, the predicted delta and the most recent lag (this model is denoted by a double asterisk in Table 2 - **). 2. Experimental Details Each data set was partitioned into consecutive segments of 75% (training data) 15% (test data) and 1% (validation data). The Neural Network model weights are adjusted using the Backpropagation algorithm on the training set and stopped (using the stopping method described later) using the root mean square errors on the training set and test set. In the case of, four different models with pattern sizes of 2,3,4,and 5 were fitted to the training set and compared on the test set data. The best performing model on the test set was then chosen for use on the validation set and as the input to the - model. The best performing model was judged as the lagged model which had the highest number of best error statistics out of the six statistics used: R 2, Percentage direction success, Root Mean Square Error (RMSE), Geometric Mean Relative Absolute Error (GMRAE), Mean Average Percentage Error (MAPE) and Percentage better than random walk (BRW). When two models performed equally well using this criteria, two - models were fitted. As the model needs a store of historical data before it can make predictions (and therefore be of use as an input to a non-linear system such as a Neural Network), the data split is slightly different for the - models. The end 1% of the data is again set aside as an unseen validation set, and of the remaining 9% data, the first 4% is used as the estimation data for the model to be fitted to. The remaining 5% is split such that the neural network is fitted to the next 4% and tested on the following 1%. This results in a 4/4/1/1 split as opposed to the 75/15/1 split used in the other models.

3. Results The results for the performance of the two hybrid models, and Neural networks is shown in Table 2 on a range of error measures recommended by Armstrong and Collopy [1]. It is evident that no single model performs the best on all error measures. In this table, we have picked the best performing models on the basis of how it performs on test data. For, we vary the parameter k (pattern size for matching) between 2 and 5 and select those models that perform the best on most number of error measures. In some cases, more than one model is chosen as the best performer as two models can perform equally well on different error measures. Out of the two hybrid models, we find that * model consistently outperforms the second model and therefore we use this as our base hybrid model for comparison with other methods. It is difficult to visualise an overall winner on each measure. In order to interpret results we simplify this process in Figure 1. We find that on most time-series, the hybrid model is the best for generating the lowest GMRAE error and is significantly good on generating high direction rate success. It outperforms the two other models on all error measures in a small proportion of cases, and there is no generic trend to comment on. From these experiments it is abundantly clear that the hybrid model has a considerable future in the forecasting domain. Some of the improvement in results is small, however any improvement is of considerable importance in financial domains, where the capital involved can be extremely large. 4. Conclusions In this paper we have proposed a novel method of combing pattern matching techniques with neural networks. This hybrid strategy has the advantage that it becomes possible to use historic information efficiently, which is not possible in traditional neural network models. In general, the hybrid model is a better choice in situations where the model selection process relies heavily on low error on GMRAE and high direction success. On other error measures, the hybrid model does not, in the majority of time series, outperform the and neural net stand-alone models. Our understanding is that the hybrid strategy is of considerable use in a range of prediction domains and it is only through empirical analysis, we can interpret its advantages. References 1. Armstrong S. and Collopy F. Error measures for generalizing about forecasting methods: Empirical comparisons. International Journal of Forecasting 1992; 8:69-8. 2. Azoff M.E. Neural Network Time Series Forecasting of Financial Markets. John Wiley, 1994. 3. Gately E. Neural Networks for Financial Forecasting. John Wiley, 1996. 4. Peters E. Fractal Market Hypothesis: Applying Chaos Theory to Investment and Economics. John Wiley, 1994.

5. Refenes A.N., Burgess N. and Bentz Y. Neural networks in financial engineering: A study in methodology. IEEE Transactions on Neural Networks 1997; 8:6:1222-1267. 6. Singh S. A long memory pattern modelling and recognition system for financial timeseries forecasting. Pattern Analysis and Applications 1999; 2:3:264-273. 7. Singh S. and Fieldsend J.E. Financial Time Series Forecasts using Fuzzy and Long Memory Pattern Recognition Systems. IEEE International Conference on Computational Intelligence for Financial Engineering, New York, (26-28 March, 2). IEEE Press, 2. Data set Description Observations A Laser generated data 1 B1, B2, B3 Physiological data, spaced by.5 second intervals. 34 each The first set is the heart rate, the second is the chest volume (respiration force), and the third is the blood oxygen concentration (measured by ear oximetry). C tickwise bids for the exchange rate from Swiss francs 3 to US dollars; they were recorded by a currency trading group from August 7, 199 to April 18, 1991 D1, D2 Computer generated series 5 E Astrophysical data. This is a set of measurements of 2724 the light curve (time variation of the intensity) of the variable white dwarf star PG1159-35 during March 1989, 1 second intervals. BRL Daily Exchange Rate of Brazilian Reals to the US 1747 Dollar (4 & ¾ years) 22 nd Oct 95 to 2 nd August 2 CAN Daily Exchange Rate of Canadian Dollars to the US 2722 Dollar (8 years) 3 rd August 1992 to 2 nd August 2 DEM Daily Exchange Rate of German Marks to the US 2722 Dollar (8 years) 3 rd August 1992 to 2 nd August 2 JPY Daily Exchange Rate of Japanese Yen to the US 2722 Dollar (8 years) 3 rd August 1992 to 2 nd August 2 CHF Daily Exchange Rate of Swiss Francs to the US 2722 Dollar (8 years) 3 rd August 1992 to 2 nd August 2 GBP Daily Exchange Rate of Pounds Sterling to the US Dollar (8 years) 3 rd August 1992 to 2 nd August 2 2722 Table 1. 1991 Santa Fe Competition data & exchange rate data Exchange Rates are daily average Interbank rates (where the average is calculated as the mid point of the low and the high of that day), 14 Data sets in all, longest being 5, points, shortest 1, points.

Dataset Model k R 2 % Dir. Suc RMSE GMRAE ( 1-3 ) MAPE BRW A 4.98326 97..98933 1.635 1.28 95. (bp) -.9977 94.949.39313 1.2764 7.41 92.929 nn-pmrs* 4.9845 93.137.86214 1.6978 11.287 93.137 4.9826 95.98.9955 1.7974 12.146 88.235 B1 2.99366 7.647.77997.29556 11.24 35.794 5.98997 63.471.9866.26968 14.356 44.735 (bp) -.9824 78.464.13187.25546 847.26 38.364 nn-pmrs* 2.9711 78.182.15517.25512 111 47.636 nn-pmrs* 5.9455 76.125.21848.2678 1446.6 5.346 2.98657 75.577.1449.2639 646.56 22.45 5.97112 76.9.15474.26463 17.6 38.322 B2 2.83335 6.147 72.295.522 67.339 41.647 5.7773 67.412 9.893.44723 81.976 49.324 (bp) -.9151 7.315 44.175.847 58.31 56.899 nn-pmrs* 2.947 65.542 44.439 1.999 59.67 52.422 nn-pmrs* 5.91469 67.734 42.434.88111 59.27 53.749 2.89625 6.438 46.421.94134 64.565 44.867 5.928 64.475 44.652.95816 61.165 5.865 B3 3.98993 51.29 21.219.438 1.3969 3.882 4.999 54.529 21.52.488 1.3522 33.735 5.98993 54.647 21.229.39971 1.422 35.588 (bp) -.9866 55.75 23.318 4.163 4.6472 51.368 nn-pmrs* 3.8264 49.683 69.114 1.9272 7.1579 41.522 nn-pmrs* 4.9891 65.484 22.449 1.49 2.539 6.467 nn-pmrs* 5.9891 65.484 22.449 1.49 2.539 6.467 49.51 3.2197 69.55 5.9163.239 4.51 1 13 4.994 48.875 2.42 1.2641 2.1728 43.8 5.9933 48.32 2.25 1.467 2.2562 42.762 C 2 1. 47.8 2.15 1-5.1919.485 36.333 (bp) -.9995 43.481.524.19 1.9894 43.481 nn-pmrs* 2.9982 43.399.873.18932 3.3581 43.431 2.99829 43.399.854.18932 3.2846 43.431

Dataset Model k R 2 % Dir. Suc RMSE GMRAE ( 1-3 ) MAPE BRW D1 4.98671 77.66.931.19888 1.457 63.4 5.98787 79.54.894.19884 1.69 66.12 (bp) -.9964 86.737.489.199 6.396 7.454 nn-pmrs* 4.98973 8.84.79.19485 8.3753 64.84 nn-pmrs* 5.99416 85.275.549.19477 6.757 69.647 4.9924 81.961.647.19486 7.716 61.941 5.99287 82.627.617.19483 7.5616 63.882 D2 4.98535 79.68.929.19885 11.869 64.5 (bp) -.9923 83.957.644.199 8.872 65.913 nn-pmrs* 4.9944 86.863.542.19469 8.282 72.314 4.99173 83.98.654.19481 9.2745 63.49 E 4.57996 56.487.3196.36995 7.929 1 17 41.455 BRL CAD CHF DEM (bp) -.596 67.929.1641.366 2.76 39.757 1 13 nn-pmrs* 4.58428 67.916.1628.35847 2.343 4.159 1 13 4.5136 65.321.1759.3589 2.283 1 13 35.4 2.99988 56..1336 2.8258.516 17.143 (bp) -.99994 46.552.943 2.8445.57 33.333 nn-pmrs* 2.99989 64.423.169 4.7359.882 39.423 2.99989 64.423.1656 4.7359.8581 39.423 2.99998 47.44.324 1.911.2729 27.645 (bp) -.99998 54.983.328 1.9211.3115 4.893 nn-pmrs* 2.99998 54.27.352 1.89.3411 42.953 2.99998 54.362.354 1.89.3425 42.617 2.99991 48.85.817 1.9384.6265 27.645 (bp) -.99856 44.674.3165 1.9493 2.8534 43.299 nn-pmrs* 2.9979 41.275.3481 1.8639 3.2294 4.64 2.9976 42.282.3727 1.8638 3.465 4.94 2.99993 5.853.898 1.6618.5617 26.621 (bp) -.9973 4.55.5284 1.6771 3.7971 4.55 nn-pmrs* 2.99622 4.64.56 1.664 4.1274 4.64 2.99641 41.611.5539 1.662 3.9973 41.611

Dataset Model k R 2 % Dir. Suc GBP JPY RMSE GMRAE ( 1-3 ) MAPE BRW 2.99995 52.218.229 2.9284.4277 29.693 (bp) -.99997 51.89.189 2.9496.385 3.584 nn-pmrs* 2.99996 45.973.199 2.8721.3839 29.866 2.99995 46.39.26 2.8721.3934 29.866 2.99988 47.44.6916 3.733.6552 24.915 (bp) -.99993 51.89.4711 3.657.5727 31.271 nn-pmrs* 2.99987 55.34.58962 3.764.7538 37.919 2.99985 56.376.63695 3.787.8195 38.255 Table 2. Table of results for each time series. The results are on the performance of the chosen model, model and the two fitted models using inputs on the test set. Results in bold indicate the best result for that error term. * 13% 38% * 29% 35% 49% 36% 1 (a) Highest R-Squared value 1 (b) Highest direction success * 21% 21% 43% 36% * 65% 14% 1 (c) Lowest RMSE 1 (d) Lowest GMRAE

* 7% 14% 43% 5% * 5% 36% 1 (e) Lowest MAPE 1 (f) Highest BRW Figure 1 (a)-(f) Proportion of models with the best performance on the error measures averaged across all data tested. APPENDIX: DATA SETS 25 2 1 8 15 6 1 4 5 2 3 6 9 12 151 181 211 241 271 32 332 362 392 423 453 483 513 543 574 64 634 664 695 725 755 785 815 846 876 96 936 966 997 997 225 353 482 511 6138 7167 8195 9223 1252 1128 1238 13337 14365 15393 16422 1745 18478 1957 2535 21563 22592 2362 24648 25676 2675 27733 28761 2979 3818 31846 32875 3393 Data Set A Data Set B1 3, 2, 1, -1, -2, -3, 5, -5, -1, -15, -2, -25, 134 21 3167 4234 53 6367 7434 85 9567 1634 117 12767 13833 149 15967 1733 181 19167 2233 213 22367 23433 245 25566 26633 277 28766 29833 39 31966 3333 134 21 3167 4234 53 6367 7434 85 9567 1634 117 12767 13833 149 15967 1733 181 19167 2233 213 22367 23433 245 25566 26633 277 28766 29833 39 31966 3333 Data Set B2 Data Set B3 1.44 1.42 1.4 1.38 1.36 1.34 1.32 1.3 1.28 1.26 1.24 1 885 1797 271 3622 4535 5447 636 7272 8185 997 11 1922 11835 12747 1366 14572 15485 16397 1731 18222 19135 247 296 21872 22785 23697 2461 25522 26435 27347 2826 29173 1433 2912 439 5869 7348 8827 135 11784 13263 14741 1622 17699 19178 2656 22135 23614 2592 26571 285 29529 317 32486 33965 35443 36922 3841 3988 41358 42837 44316 45794 47273 48752 Data Set C Data Set D1

1433 2912 439 5869 7348 8827 135 11784 13263 14741 1622 17699 19178 2656 22135 23614 2592 26571 285 29529 317 32486 33965 35443 36922 3841 3988 41358 42837 44316 45794 47273 48752 1 78 1584 2389 3193 3998 483 567 6412 7216 821 8825 963 1434 11239 1243 12848 13652 14457 15261 1666 1687 17675 18479 19284 288 2893 21698 2252 2337 24111 24916 2572 26525 Data Set D2 Data Set E 51 12 154 25 257 39 36 412 464 515 567 619 67 722 773 825 877 928 98 132 183 1135 1187 1238 129 1341 1393 1445 1496 1548 16 1651 173 2 1 87 175 264 353 442 531 62 79 797 886 975 164 1153 1242 133 1419 158 1597 1686 1775 1864 1952 241 213 2219 238 2397 2486 2574 2663 2752 2841 1.55 1.5 1.45 1.4 1.35 1.3 1.25 1.2 Data Set BRL Data Set CAD 86 173 261 349 437 525 613 71 788 876 964 152 114 1228 1315 143 1491 1579 1667 1755 1843 193 218 216 2194 2282 237 2458 2545 2633 2721 289 2897 1.7 1.6 1.5 1.4 1.3 1.2 86 173 261 349 437 525 613 71 788 876 964 152 114 1228 1315 143 1491 1579 1667 1755 1843 193 218 216 2194 2282 237 2458 2545 2633 2721 289 2897 2.1 2 1.9 1.8 1.7 1.6 1.5 1.4 Data Set CHF Data Set DEM 87 175 264 353 442 531 62 79 797 886 975 164 1153 1242 133 1419 158 1597 1686 1775 1864 1952 241 213 2219 238 2397 2486 2574 2663 2752 2841.7.68.66.64.62.6.58.56.54.52.5 86 174 263 351 439 528 616 75 793 881 97 158 1146 1235 1323 1411 15 1588 1676 1765 1853 1941 23 2118 226 2295 2383 2471 256 2648 2737 2825 2913 14 13 12 11 1 9 Data Set GBP Data Set JPY