TIME SERIES FORECASTING WITH MULTIPLE CANDIDATE MODELS: SELECTING OR COMBINING?

Size: px
Start display at page:

Download "TIME SERIES FORECASTING WITH MULTIPLE CANDIDATE MODELS: SELECTING OR COMBINING?"

Transcription

1 Vol. 18 No. 1 Journal of Systems Science and Complexity Jan., 2005 TIME SERIES FORECASTING WITH MULTIPLE CANDIDATE MODELS: SELECTING OR COMBINING? YU Lean (Institute of Systems Science, Academy of Mathematics and Systems Science, Chinese Academy of Sciences, Beijing , China; School of Management, Graduate School of Chinese Academy of Sciences, Beijing , China) WANG Shouyang (Institute of Systems Science, Academy of Mathematics and Systems Science, Chinese Academy of Sciences, Beijing , China; Institute of Policy and Planning Sciences, University of Tsukuba, Tsukuba, Ibaraki , Japan. sywang@amss.ac.cn) K. K. Lai (Department of Management Sciences, City University of Hong Kong, Tat Chee Avenue, Kowloon, Hong Kong) Y. Nakamori (School of Knowledge Science, Japan Advanced Institute of Science and Technology 1-1, Asahidai, Ishikawa , Japan) Abstract. Various mathematical models have been commonly used in time series analysis and forecasting. In these processes, academic researchers and business practitioners often come up against two important problems. One is whether to select an appropriate modeling approach for prediction purposes or to combine these different individual approaches into a single forecast for the different/dissimilar modeling approaches. Another is whether to select the best candidate model for forecasting or to mix the various candidate models with different parameters into a new forecast for the same/similar modeling approaches. In this study, we propose a set of computational procedures to solve the above two issues via two judgmental criteria. Meanwhile, in view of the problems presented in the literature, a novel modeling technique is also proposed to overcome the drawbacks of existing combined forecasting methods. To verify the efficiency and reliability of the proposed procedure and modeling technique, the simulations and real data examples are conducted in this study. The results obtained reveal that the proposed procedure and modeling technique can be used as a feasible solution for time series forecasting with multiple candidate models. Key words. forecasts. Time series forecasting, model selection, stability, robustness, combining Received July 20, *This paper was partially supported by NSFC, CAS, RGC of Hong Kong and Ministry of Education and Technology of Japan.

2 2 YU LEAN et al. Vol Introduction With the increasing advancement of computational technology, a variety of models (both linear and nonlinear) have been widely used for time series modeling and forecasting. For example, the autoregressive integrated moving average (ARIMA) model proposed by Box and Jenkins [1], which is a typical linear model, has been proven to be effective in many time series forecasts. Likewise, the artificial neural network (ANN) model, a typical nonlinear model and newly intelligent computational technique, has also been shown to be a very promising approach for time series modeling and forecasting. Due to the fact that different time series have different properties and features, the selected model based on a selection criterion (e.g., Akaike information criterion (AIC) [2] ), hypothesis testing and graphical inspections, is often mis-specified and thus may cause an unexpectedly high variability in the final prediction and affect forecasting accuracy and reliability. In practical application, however, academic researchers and business practitioners often face two important dilemmas. One is whether to select an appropriate modeling approach for prediction purposes or to combine these different individual approaches into a single forecast for dissimilar modeling approaches. The other is whether to select the best candidate model to forecast or to mix the various candidates with different parameters into a new forecast for the same/similar modeling approaches. In other words, for time series forecasting the focus is on whether to select a model from multiple candidates or to combine these different models into a single forecast. Generally speaking, the two issues are highly non-trivial and have received considerable attention with different approaches being studied. Here we briefly discuss some of these approaches that are closely related to our work. In time series forecasting applications, the general practice is to select one appropriate model from the multiple candidates in terms of some selection methods; thus final estimation, interpretation, and prediction are then based on the selected model. Generally speaking, there are three methods for multiple candidate model selection. The first is graphical inspection together with examination of simple summary statistics (such as autocorrelations (AC) and partial autocorrelations (PAC)), which is very useful for preliminary modeling analysis. The second is hypothesis testing, a formal technique for model selection in statistics. The third method is to use a well-defined and formal model selection criterion, such as AIC [2] or Bayesian information criterion (BIC) [3]. However, these methods exhibit some defects. For example, the first method is too subjective and too rough in general for model selection. Likewise, there are difficulties with the second approach due to the challenging issue of multiple testing (e.g., there is no objective guideline for the choice of the size of each individual test, and it is completely unclear how such a choice affects the forecasting accuracy). Furthermore, Breiman [4] pointed out that the estimators based on model selection are instable. Yang [5] also considered that a major drawback with model selection is its instability. For example, with a small or moderate number of observations, as expected, it is usually hard to distinguish models that are close to each other (the model selection criterion values are generally quite close). In this case, the choice of the model with the smallest criterion value is unstable. That is, a slight change in the data may lead to a different model being chosen. As a consequence, the forecasts based on the selected model are highly variable [5 6]. Furthermore, Yang [5] pointed out identifying the true model is not necessarily optimal for forecasting. To reduce variability in model selection, researchers have turned to combining or mixing the different candidate models for prediction purposes. A variety of literature on combined forecasting is reported. In the combination of the same/similar method, Draper [7] and George and McCulloch [8] proposed an averaging method to combine a stabilized estimator. Raftery [9] suggested using a BIC approximation for Bayesian model averaging. Madigan and York [10] used a Markov chain Monte Carlo approximation to obtain a stable forecast. Breiman [11] proposed a bagging method to generate multiple versions of the estimator and then average them into a

3 No. 1 TIME SERIES FORECASTING: SELECTING OR COMBINING? 3 stable estimator. In a similar manner, Buckland et al. [12] proposed a plausible model weighting method according to the value of a model selection criterion (e.g., AIC). Cross-validation and bootstrapping have also been used to linearly combine different models with the intention of improving accuracy by finding the best linear combination (see [4,13,14]). Juditsky and Nemirovski [15] proposed a stochastic approximation method to combine k forecasts in the best linear combination. Yang [6] applied the adaptive regression method (ARM) by mixing different candidates from the same model for regression. Likewise, Yang [5] used the aggregated forecast through exponential re-weighting method (AFTER) to combine the forecasts from the different individual autoregressive moving average (ARMA) models. However, their works are only limited to linear combinations of the same methods with different parameters. Readers can refer to a review of this topic by Hoeting et al. [16] for more details. In addition, more references are presented in References [17 37]. In the combination of different/dissimilar models, literature documenting the research shows it to be quite diverse. Combining forecasts with dissimilar methods has been studied for over three decades, starting with the pioneering work of Reid [38,39] and Bates and Granger [40]. Various methods have been involved. For work pre-dating 1989, readers can refer to Clemen [41] for a comprehensive review of this topic. After that time, with the increasing development and application of new computational technology, many artificial intelligent (AI) techniques such as artificial neural networks (ANNs) have been presented. There are a few examples in the existing literature combining neural network forecasting models with conventional time series forecasting techniques. For example, Wedding II and Cios [42] constructed a combination model integrated radial basis function neural networks (RBFNN) and the un-invariant Box-Jenkins (UBJ) model to predict three time series. Luxhoj et al. [43] presented a hybrid econometric and ANN approach for sales forecasting. Likewise, Voort et al. [44] introduced a hybrid method called KARIMA using a Kohonen self-organizing map and ARIMA method to predict short-term traffic flow. Recently, Tseng et al. [45] proposed a hybrid model (SARIMABP) that combines the seasonal ARIMA (SARIMA) model and the back-propagation (BP) neural network model to predict seasonal time series data. Zhang [46] applied a hybrid methodology that combined both ARIMA and ANN models for time series forecasting. For more details, readers can refer to References [17 37]. The idea of combining forecasts implicitly assumed that one could not identify the underlying process (i.e., one could not select an appropriate model for a specified time series), but that different forecasting models were able to capture different aspects of the information available for prediction. This is also one of starting points for this study. As Clemen [41] concludes: Using a combination of forecasts amounts to an admission that the forecaster is unable to build a properly specified model. To try ever more elaborate combining models seems to add insult to injury as the more complicated combinations does not generally perform all that well. Hence combining the different candidate models for prediction has been common practice in practical applications to reduce the variability and uncertainty of selecting an individual model. It should be noted that, however, two main problems are found in the existing combined forecast models mentioned previously. One is the form of the combined forecasting model. In the existing literature, the combined forecasting models are limited to linear combination form. But a linear combination approach is not necessarily appropriate for all the circumstances. Another issue is the number of individual forecasting models used. The question here is how to determine the number of individual forecasting models. It is well known that not all circumstances are satisfied by following the more, the better. Thus, it is necessary to determine the number of individual models used in the combined forecasts. With respect to the first problem (i.e. linear combination drawback), a nonlinear combined forecasting model utilizing the ANN technique is introduced in this paper. With regard to

4 4 YU LEAN et al. Vol. 18 the second problem (i.e. determining the number of individual models used in combined forecasting), the principal component analysis (PCA) technique is introduced. We use the PCA technique to choose the appropriate number of individual models. Therefore this study proposes a novel nonlinear combined forecasting model utilizing the ANN technique and PCA technique to achieve a reduction in model uncertainty, or achieving an increase in forecast accuracy. The PCA & ANN-based nonlinearly combined (PANC) forecasting approach is used in the proposed novel nonlinear combined model. However, the fact that we use the combined method for time series prediction does not necessarily mean that we are against the practice of model selection. Generally, identifying the true model (when it makes good sense) is an important task in understanding relationships of time series variables. In deterministic linear time series, it is often observed that selection may outperform combining methods when one model is very strongly preferred, in which case there is little instability in selection. As Yang [5] claimed, in the time series context, combining does not necessarily lead to prediction improvement when model selection is stable. That is to say, the use of combining forecasts does not mean that the individual forecasts should be suppressed. In fact, these individual forecasts can provide evidence about the relative merits of their respective models, methods, or forecasters as well as about the relationships among the multiple forecasts. Furthermore, for purpose of reporting and comparison, it is helpful to give the individual forecast as well as the combined forecast in order to provide an indication of the variability among the forecasts. Decision makers using the forecasts may want to see if and how their decisions change when different forecasts are used [47]. In addition, it should be pointed out the combining approach we take is related to (although different from) formal Bayesian consideration. In particular, no prior distribution will be specified for parameters in the models. The aims of this study, then, are three-fold: (1) to reveal how to deal with the two dilemmas of selecting and combining; (2) to show how to construct a novel nonlinear combination model for time series forecasting; and (3) to compare the accuracy of various methods and models with real data examples. In view of the three aims, this study firstly proposes a double-phaseprocessing procedure to treat the two dilemmas of selecting or combining, moves on to describe the building process for the proposed nonlinear combination model, and finally presents the application of the proposed procedure and nonlinear combined approach with the real data examples. The rest of the study is organized as follows. Section 2 describes the procedure for dealing with the two dilemmas of selecting and combining in detail. In Section 3, a novel nonlinear combined forecasting model is proposed to overcome the two problems of existing combined forecasts. In order to verify the effectiveness and efficiency of the proposed procedure and model, simulation experiments and empirical analysis of real time series are reported in Section 4. Conclusions and future research directions are given in Section 5. 2 Dilemmas of Selecting or Combining and Solutions to Them Assume {y 1, y 2,, y t } to be a time series. At time t for t > 1, we are interested in forecasting the next value y t+1 based on the past observations of y 1, y 2,, y t. Here we focus on one-step-ahead point forecasting. As mentioned previously, two dilemmas are taken into account in this study. With respect to the same time series, the first is whether to select a true model or to combine multiple individual models for the different/dissimilar methods; the other is whether to select a true model or to combine multiple models with different parameters for the same/similar method. That is, one faces two types of choice for a specified time series. If we do not decide on a

5 No. 1 TIME SERIES FORECASTING: SELECTING OR COMBINING? 5 determined method for the time series, then multiple different/dissimilar methods are tested by trial and error: thus one faces the first choice. If a method is confirmed that is appropriate for the time series, multiple versions with different parameters of the determined method are generated: i.e., one faces the second choice. This is graphically represented in Fig.1. From Fig. 1, we can clearly see the two dilemmas for time series forecasting. In short, this is a problem of selecting or combining. Subsequently, we further explore the solution of the two dilemmas. Many selection methods are often utilized in solving the selecting or combining problem for multiple candidate models. However, as mentioned earlier, the potentially large variability in the resulting forecast is common to all model selection methods. When there is substantial uncertainty in finding the best model, alternative methods such as combination model should be taken into account. An interesting issue of selecting or combining is: how should we measure the model selection criteria or the resulting prediction? This is the key to solve the selecting or combining dilemma. It should be noted that two judgmental criteria are required to define dealing with two types of dilemma. In order to solve the dilemmas, a double-phase-processing procedure is proposed. Time series Different methods Same method Model (A) Model (B) Model (K) Model (A 1) Model (A 2) Model (A n) Selection or combining (Dilemma I) Judgmental criterion I Decision I Selection or combining (Dilemma II) Judgmental criterion II Decision II Fig. 1 Two dilemmas of time series forecasting In the first phase, a solution to the first dilemma is provided. In the case of whether to select or combine the different/dissimilar methods, the relative prediction performance stability (S) is introduced. Here the conception of prediction performance stability is that time series prediction performance with different/dissimilar model is stable. In other words, for a specified time series, if the difference between the performance indicator for the best candidate model and the performance indicator for the remaining models is very large, then the prediction performance for the time series is unstable and another strategy, such as combining method, should be considered. Otherwise, the time series prediction performance is stable and selecting the best model from the different/dissimilar candidates may be a wise decision. To be concrete, assume that a different model {A, B,, K} is determined based on all the observations {y i } (i = 1, 2,, k) and use them as predictors. Accordingly, in the same prediction period, the prediction performance indicator (e.g., root mean squared error (RMSE)), defined by {P I i } (i = 1, 2,, k) and its volatilities (σ i ) (i = 1, 2,, k) with the different evaluation

6 6 YU LEAN et al. Vol. 18 sets, are calculated. Hence, the relative prediction performance stability (S) is calculated within the framework of the different evaluation sets. S i = ( 1 m m j=1 P I ij)/σ i, i = 1, 2,, k, (1) Min {P I i }/σ P Ii i where P I i = m j=1 P I ij m. By comparing S i and S θ, we can judge whether to select a true model or combine multiple different/dissimilar models, i.e., Decision = { Selecting, if Si S θ, Combing, if S i < S θ. i = 1, 2,, k. (2) From Equation (2), we can see that if the all the S i are larger than the threshold value S θ, a model with the smallest value of P I is selected. If the opposite is true, the combining strategy is recommended. In other words, once the value of S θ is determined, we can make corresponding decisions for the first dilemma. It should be noted that it is hard to determine a rational value of S θ. Furthermore, once S θ is determined, we face a probable situation: some values of S i are larger than S θ and others smaller than S θ. If this is the case, how to deal with the situation is still a problem. When the selecting decision is made in the first phase, the second dilemma is generated. In the second phase, we use the robustness judgmental criterion. When the robustness of the selected model based on some model selection criteria is strong (i.e., when one model is strongly preferred in terms of the corresponding selection criterion), the model should be selected rather than combining from the multiple candidate models with different parameters for the same method. On the other hand, when the robustness of the selected model is weak, model combining should be considered in order to obtain good prediction performance. Here the definition of model robustness has been a focus issue. In fact, model robustness can be defined in such a way as follows. The selected model should be robust in the sense that it is indifferent to a radical change to a small portion of the data or a slight change in all of the data [48]. Simply speaking, if a model is robust, a minor change in the data should not change the outcome dramatically. The idea of measuring model robustness is to examine its consistency with different data sizes. Suppose that the model M n is selected by the model selection criterion based on the observations {y i } (i = 1, 2,, n). Let k be an integer between 1 and n 1. For each j in {n k, n k +1,, n 1}, apply the model selection method to the data {y i } (i = 1, 2,, j) and let M j denote the selected model. Then let R (robustness metric) be the percentage of time series for which the same model ( M n ) is selected, i.e., R = n 1 j=n k I{ M j = M n } k, (3) where I{ } denotes the indicator function: that is, if M j = M n, then I{ M j = M n } = 1, else I{ M j = M n } = 0. The rationale behind the consideration of R is quite clear: removing a few observations should not cause much change for a stable model. In addition, the integer k should be chosen appropriately. On one hand, we want to choose small k so that the selection problems

7 No. 1 TIME SERIES FORECASTING: SELECTING OR COMBINING? 7 for j in {n k, n k+1,, n 1} are similar to the real problem with full observations. On the other hand, we want k not to be too small so that R is reasonably stable [6]. Once the rational threshold value R θ is determined, we can make the decision of whether to select or combine by judging the robustness of the selected model in the second dilemma, i.e., { Selecting, if R Rθ, Decision = (4) Combing, if R < R θ. That is, if the robustness value of the selected model is larger than the predefined threshold value, then the selecting strategy is preferred; else the combining strategy is preferred. Although Yang [6] also pointed out the same question, his work was only limited to the selection of the same method. Thus, the dilemmas we present are wider and the alternative solution is more general. Therefore, the proposed procedure is different from previous work. To summarize, the proposed double-phase-processing procedure is executed as follows. In the first sub-procedure, we use the performance stability indictor to deal with the first dilemma (Equations (1) (2)). In the second sub-procedure, we use the robustness metric to treat the second dilemma, as Equations (3) (4) shows. With regard to the specified time series, two dilemmas are often faced. With respect to the specified time series, common practice is to use multiple dissimilar methods to model the problem by trial and error and then introduce the judgmental criteria to make the corresponding decisions. That is, for many different/dissimilar models, if the time series model has stable performance, then a selecting decision is made, otherwise the combining strategy is recommended. If the model is robust, then we select a best model from the multiple candidates with different parameters, otherwise we combine the candidates into a single time series forecast. A flow chart of proposed procedure is depicted in Fig. 2. Time series data Multiple different candidate models with dissimilar method Stability No Yes Selecting one method Multiple candidate models with different parameters for name method Yes Selecting Robustness No Combining Selection Criterion Forecasting Combined Strategy Fig. 2 The double-phase-processing procedure of time series forecasting

8 8 YU LEAN et al. Vol. 18 Relying on the proposed double-phase-processing procedure, the above two types of dilemma can be solved. This gives a clear decision on whether to select an appropriate model from multiple candidates or to combine these models for a specified time series. In practical application, however, the combining strategy is often selected due to the uncertainty of time series formation. Thus one has to confront combination forecast problems. As mentioned in Section 1, several drawbacks of combination models are found in the existing literature. Subsequently, a novel nonlinear combination model is proposed in the following section to overcome the shortcomings of existing combination methods. 3 Nonlinear Combination Forecasting Model The idea of a combined forecasting model is not new. Since the pioneering work of Reid [38 39] and Bates and Granger [40], a variety of studies have been conducted and many combined methods have been proposed. Basically, the main problem of combined forecasts can be described as follows. Suppose there are n forecasts such as ŷ 1 (t), ŷ 2 (t),, ŷ n (t) (including similar and dissimilar forecasts). The question is how to combine these different forecasts into a single forecast ŷ(t), which is assumed to be a more accurate forecast. The general form of the model for such a combined forecast can be defined as: n ŷ(t) = w i ŷ i (t), (5) i=1 where w i denotes the assigned weight of ŷ i (t), and in general the sum of the weights is equal to 1, i.e., w i = 1. In some situations the weights may have negative values and the sum of i them may be greater than 1 [49]. There are a variety of methods available to determine the weights used in the combined forecasts. For comparison, three main methods are presented here. The first is the equal weights (EW) method, which uses a simple arithmetic average of the individual forecasts, and which is a relatively easy and robust method. That is, w i = 1 n, (6) where n is the number of forecasts. The second is minimum-variance (MV) method proposed by Dickinson [50 51]. Its main ideas are Min(w i V wi T), n (P 1 ) w i = 1, i = 1, 2,, n, (7) i=1 w i 0, where V is the matrix of error-variance. By solving the quadratic programming (P 1 ), an optimal weight set can be obtained for the combining forecasts. The third is the time-varying minimum-square-error (TVMSE) method proposed by Yu et al. [52] and its main ideas are as Min(w i (e i (t)) 2 ), n (P 1 ) w i = 1, i = 1, 2,, n; t = 1, 2,, N, (8) i=1 w i 0,

9 No. 1 TIME SERIES FORECASTING: SELECTING OR COMBINING? 9 where t represents the evaluation/prediction period. With the change of time t, the weights vary with time t. This method is thus a new and robust method for combining forecasts. (Refer to Clemen [41] and References for more methods.) However, there are two main problems in existing combined forecasting approaches. One is the number of individual forecasts. Theoretical proof [53] shows that the total forecasting errors of combined forecasts does not necessarily decrease with an increase of the number of individual forecasting models. If this is the case, there is redundant or repeated (even noisy) information among the individual forecasts. Furthermore, the redundant information often directly impacts on the effectiveness of the combined forecasting method. It is therefore necessary to eliminate individual forecasts that are not required. In other words, one of the key problems of combined forecasts is to determine a reasonable number of individual forecasts. The other problem is that the relationships among individual forecasts are determined in a linear manner. The three methods mentioned above, and many other weighted methods, are all developed under linear consideration. A combined forecast should merge the individual forecasts according to the natural relationships existing between the individual forecasts including, but not limited to, linear relationships. Because a linear combination of existing information cannot represent the relationship between individual forecasting models in some situations, it is necessary to introduce nonlinear combination methodology in the combined forecasts. We set out to solve the two problems in the following. In the first problem, the question we face is how to extract effective information that reflects substantial characteristics of series from all selected individual forecasts and how to eliminate redundant information. The principal component analysis (PCA) technique (see [54,55]) is used as an alternative tool to solve the problems. The PCA technique, an effective feature extraction method, is widely used in signal processing, statistics and neural computing. The basic idea in PCA is to find the components (s 1, s 2,, s p ) that can explain the maximum amount of variance possible by p linearly transformed components from a data vector with q dimensions. The mathematical technique used in PCA is called eigen analysis. In addition, the basic goal in PCA is to reduce the number of dimensions of the data. Thus, one usually chooses p q. Indeed, it can be proved that the representation given by PCA is an optimal linear dimension reduction technique in the mean-square sense [54]. There are important benefits to such a reduction in dimensions. First, the computation of the subsequent processing is reduced. Second, noise can be reduced and the meaningful underlying information can be identified. The following presents the PCA process for combined forecasting. Assuming that there are n individual forecasting models in the combined forecasts and that every forecasting model contains m forecasting results, then forecasting matrix (Y ) can represented as y 11 y 12 y 1m y 21 y 22 y 2m Y =......, (9) y n1 y n2 y nm where y ij is the jth forecasting value with the ith forecasting model. Next, we deal with the forecasting matrix using the PCA technique. First, eigen-values (λ 1, λ 2,, λ n ) and corresponding eigenvectors A = (a 1, a 2,, a n ) can be solved from the forecasting matrix. Then the new principal components are calculated as Z i = a T i Y, i = 1, 2,, n. (10)

10 10 YU LEAN et al. Vol. 18 Subsequently, we choose m (m n) principal components from existing n components. If this is the case, the saved information content is judged by θ = (λ 1 + λ λ m ) (λ 1 + λ λ n ). (11) If θ is sufficiently large (e.g., θ > 0.8), enough information has been saved after the feature extraction process of PCA. Thus, re-combining the new information can further improve the prediction performance of combined forecasts. In the case of the second problem, we propose a nonlinear combined forecasting model as a remedy. The detailed model is presented as follows. A nonlinear combined forecasting model can be viewed as a nonlinear information processing system that can be represented as y = f(i 1, I 2,, I n ), (12) where f( ) is a nonlinear function and I i denotes the information provided by individual forecasting models. If an individual forecasting model can provide an individual forecast ŷ i, then Equation (12) can be represented as ŷ = f(ŷ 1, ŷ 2,, ŷ n ). (13) Determining the function f( ) is quite challenging. In this study, ANN is employed to realize nonlinear mapping. The ability of back-propagation neural networks to represent nonlinear models has been tested by previous work. That is, all individual forecasts are used as inputs of the ANN model, and the output of the ANN model is seen as the results of nonlinear combination forecasts. In fact, the ANN training is a process of searching for optimal weights. That is, this training process makes the sum of the square errors minimal, i.e., Min[(y ŷ)(y ŷ) T ], (14) or Min{[y f(ŷ 1, ŷ 2,, ŷ n )][y f(ŷ 1, ŷ 2,, ŷ n )] T }. (15) By solving the unconstrained nonlinear programming problems of linear combination models with an ANN model, a PCA & ANN-based nonlinear combination (PANC) model is generated for nonlinear combining forecasting. Although Shi et al. [49] proposed a nonlinear combined forecasting method they only solved the second problem of existing combined methods. Therefore, the PANC model is a fully novel method. To summarize, the proposed nonlinear combination model consists of four stages. Generally speaking, in the first stage we construct an original forecasting matrix according to the selected individual forecasts. In the second stage, the PCA technique is used to deal with the original forecasting matrix and a new forecasting matrix is obtained. In the third stage, based upon the judgments of PCA results, the number of individual forecasting models is determined. And in the final stage, an ANN model is developed to combine different individual forecasts; meanwhile the corresponding forecasting results are obtained. To verify the efficiency of the proposed procedures and models, several real data examples are conducted in the following section.

11 No. 1 TIME SERIES FORECASTING: SELECTING OR COMBINING? 11 4 A Simulation Study 4.1 Experimental Plan, Data Description and Forecasting Evaluation Criterion In this study, many experiments are conducted to verify the efficiency of the proposed procedure and nonlinear combined forecasting model, and two real time series are used. One is housing sales time series, which is used to check the efficiency of the proposed procedure. The other is an exchange rate series U.S dollars (USD) against British pound (GBP) which is used to test the reliability and effectiveness of the proposed procedure and of the nonlinear combined model at the same time. In this study, four forecasting methods the ARIMA model, exponential smoothing (ES) model, simple moving average (SMA) model and artificial neural network (ANN) model with three-layer network structures are used in the experiments. The housing sales data used are monthly and are obtained from Pankratz book [56], covering the period from January 1965 to December 1975 with a total of n = 132 observations. We take the monthly data from January 1965 to June 1973 as the in-sample (training periods) data sets (102 observations) and the remainder as the out-of-sample (testing periods) data sets (30 observations). The foreign exchange (USD/GBP) data used in this paper are monthly and are obtained from Pacific Exchange Rate Service ( provided by Professor Werner Antweiler, University of British Columbia, Vancouver, Canada. We take monthly data from January 1971 to December 1998 as in-sample data sets (336 observations) and take the data from January 1999 to December 2003 as out-of-sample data sets (60 observations), which are used to evaluate the good or bad performance of predictions based on evaluation measurements. In order to save space, the original data are not listed here, and detailed data can be obtained from the website or from the authors. In order to compare the prediction performance, it is necessary to introduce the forecasting evaluation criterion. It is known that one of the most important forecasting evaluation criteria is the root mean square error (RMSE). Its computational form can be defined as RMSE = 1 N (y t ŷ t ) N 2, (16) where y t (t = 1, 2,, N) represent the actual values, ŷ t (t = 1, 2,, N) are predicted values, and N is the number of evaluation periods. In this study we use the RMSE to measure the forecasting performance of different models. 4.2 Empirical Results In this study, two classes of experiment are conducted to verify the effectiveness of efficiency of the proposed procedure and nonlinear model. t= Experiments for the proposed procedure In accordance with the procedure proposed in Section 2, we use the housing sales series and exchange rate series to verify and test the proposed procedure. Some different forecasting methods are used to fit the specified time series, thus facing the first dilemma (whether to select one method from the alternatives or whether to combine these candidates into a single forecast). If the selection strategy is preferred, then the second dilemma is generated (whether to select one model from multiple candidate models with different parameters or whether to combine these candidate models with different parameters for prediction purposes). As mentioned in experimental plans, the ARIMA, ES, SMA and ANN models are used to fit the time series and to select an appropriate model for every method. Then we use the

12 12 YU LEAN et al. Vol. 18 selected model to predict the future value of time series. For convenience, the testing sets (or evaluation sets) are presented for verification. Accordingly, the performance indicators are calculated. In this experiment, two examples are presented. In view of the judgment rule of performance stability, the corresponding decision can be made. Here we assume the threshold value S θ = 4.00 and R θ = Example I: Sales forecasting In this example, the results of sales forecasting performance indicators with different evaluation sets are shown in Table 1. Table 1 The RMSE of different methods with different evaluation sets Evaluation sets P I ij ARIMA ES SMA ANN 10 observations RMSE observations RMSE observations RMSE P I i σ i P I: performance indicator; ARIMA: autoregressive integrated moving average; ES: exponential smoothing; SMA: simple moving average; ANN: artificial neural network; RMSE: root mean square error. From Table 1, we can calculate the performance stability indicator in accordance with Equation (1). First of all, we find that the smallest value of P I i (i = 1, 2, 3, 4) is , i.e., Min {P I i } = Then, the values of S i (i = 1, 2, 3) are calculated as follows. i S 1 = / / = ; S 2 = / / = ; S 3 = / / = Because all S i > S θ = 4.00 (i = 1, 2, 3) for different evaluation sets, the selection strategy is recommended. The ANN model is selected as a true model for prediction purposes. However, if we assume S θ to be 60.00, then the combination strategy is preferred. In addition, if S θ is equal to 10.00, it is hard to judge between the two strategies and further exploration is needed. This implies that determining an appropriate S θ is extremely important. When the selection strategy is preferred, as mentioned earlier, we face the second dilemma. Here we take the ANN model as an example for further explanations. First of all, we select an appropriate ANN model based on the full observations by training and learning. Then we change the sample sizes and fit the series again and again. Meanwhile, the robustness metric (R) is calculated. Finally, the corresponding decision can be made by comparing the value of R with the threshold value (R θ ). The experimental results are presented in Tables 2 3. Table 2 The process of method selection ANN(input node, hidden node, output node) Methods (3,2,1) (3,3,1) (3,4,1) (3,5,1) (4,2,1) (4,3,1) (4,4,1) (4,5,1) RMSE ANN: artificial neural network; RMSE: root mean square error.

13 No. 1 TIME SERIES FORECASTING: SELECTING OR COMBINING? 13 Table 3 The robustness testing process (k = 30) Sample size ANN(I, H, O) (3,3,1) (4,3,1) (4,4,1) (4,4,1) (4,3,1) (4,4,1) (4,4,1) (3,5,1) (4,4,1) (4,4,1) I j( Mj = Mn) Sample size ANN(I, H, O) (4,4,1) (4,5,1) (4,4,1) (3,4,1) (4,4,1) (4,4,1) (4,5,1) (4,4,1) (4,4,1) (4,4,1) I j{ Mj = Mn} Sample size ANN(I, H, O) (4,4,1) (4,4,1) (4,4,1) (4,4,1) (4,4,1) (4,4,1) (4,4,1) (4,4,1) (4,4,1) (4,4,1) I j{ Mj = Mn} If { Mj = Mn}, then I j { Mj = Mn} = 1, else I j { Mj = Mn} = 0; ANN(I, H, O): ANN (input node, hidden node, output node). From Table 2, we select the ANN(4,4,1) as an appropriate model using the corresponding model selection criteria. By changing the sample size, we can make the corresponding decisions. From Table 3, we can find that different k values often lead to different robustness metric values. When k is equal to 20, the value of R is 0.85(R > R θ ), then the decision of selecting ANN(4,4,1) as a predictor is rational. However, when k is equal to 30, the value of R is (R < R θ ), then ANN(4,4,1) should be rejected for the time series. Combining multiple models with different parameters into a new forecast is a good choice for prediction purposes. Thus, it is very important to choose k with an appropriate value, as this affects our decisions to a great extent. Unfortunately, the determination of k still depends on the trial and error. Example II: Exchange rate forecasting In this example, we use the foreign exchange series to check the efficiency of the proposed procedure. The results of exchange rate forecasting performance indicators with different evaluation sets are shown in Table 4. Table 4 The RMSE of different methods with different evaluation sets Evaluation sets P I ARIMA ES SMA ANN 30 observations RMSE observations RMSE observations RMSE P I i σ i P I: performance indicator; ARIMA: autoregressive integrated moving average; ES: exponential smoothing; SMA: simple moving average; ANN: artificial neural network; RMSE: root mean square error; σ: volatility. From Table 4, we find that the performance indicators of an ANN model with the different evaluation sets are the best of all, indicating that the ANN model is a very promising modeling technique. According to the first judgmental rule and the results in Table 4, we can calculate the performance stability indictors. For three different evaluation sets, S 1 = / / = ; S 2 = / / = ; S 3 = / / =

14 14 YU LEAN et al. Vol. 18 Because S i < S θ = 4.00(i = 1, 2, 3) for different evaluation sets, the combination strategy is recommended. However, verification of the combination forecasts is the task of the following section Experiments for the proposed nonlinear combined model As mentioned before, the combined strategy problem has to face either the first dilemma or the second dilemma. In allusion to the shortcomings of existing combined forecast model, a novel nonlinear combined model is proposed. In this section, four individual models, ARIMA, ES, SMA and ANN models, and four methods of combined forecast, EW, MV, TVMSE and ANC methods, are used for combination forecasts to verify the effectiveness of the proposed nonlinear combination method. Here, the comparison of three sides is presented. For space reasons, we only use the exchange rate series for our tests. A. Individual models vs. combination models First of all, we provide the comparison between four individual models and four combination methods. The comparison results are shown in Table 5. Table 5 The comparison between individual models and combination methods Models/Methods ARIMA ES SMA ANN EW MV TVMSE ANC RMSE Rank ARIMA: autoregressive integrated moving average; ES: exponential smoothing; SMA: simple moving average; ANN: artificial neural network; EW: equal weights; MV: minimum variance; TVMSE: timevarying minimum-squared-error; ANC: ANN-based nonlinear combination; RMSE: root mean square error. From Table 5, we can find that (1) the performance of combination methods is generally better than that of individual models; (2) among individual models, the ANN model performs the best; (3) in combination methods, the ANN-based nonlinear combination (ANC) method outperforms the three linear combination methods (EW, MV and TVMSE); and (4) the ANN-based combination method performs the best of all the individual forecast models and combination methods. B. Combination of the same/similar methods vs. combination with different/dissimilar methods Subsequently, we compare the performance between combination with same methods and combination with different methods. Here we take the ARIMA model as a proxy (the same model) for comparison. The corresponding method is presented in Table 6. Table 6 The comparison between combination with same methods and combination with different methods Methods EW ARIMA EW MV ARIMA MV TVMSE ARIMA TVMSE ANC ARIMA ANC RMSE Rank ARIMA: autoregressive integrated moving average; EW: equal weights; MV: minimum variance; TVMSE: time-varying minimum-squared-error; ANC: ANN-based nonlinear combination; RMSE: root mean square error. As can be seen in Table 6, we can conclude that (1) the performance when combining different models is better than that when combining the same method, with the exception of

15 No. 1 TIME SERIES FORECASTING: SELECTING OR COMBINING? 15 EW; and (2) according to the rank, ANC with different models performs the best, followed by MV with different models, ANC with the same method and TVMSE with different models. C. PCA-based linear (nonlinear) combination models vs. linear (nonlinear) combination models In the following, we compare the forecasting performance between combination models and PCA-based combination models, and the results are shown in Table 7. Table 7 The comparison between combination models and PCA-based combination models Models/Methods PEW EW PMV MV PTVMSE TVMSE PANC ANC RMSE Rank PEW: PCA-based equal weights; EW: equal weights; PMV: PCA-based minimum variance; MV: minimum variance; PTVMSE: PCA-based time-varying minimum-squared-error; TVMSE: time-varying minimum-squared-error; PANC: PCA & ANN-based nonlinear combination; ANC: ANN-based nonlinear combination; RMSE: root mean square error. As Table 7 reveals, we can conclude that PCA-based combination models outperform the other combination models in general in terms of RMSE and rank. Of all the combination models, the PANC performs the best, revealing that it is a robust and promising model and is worth generalization and application for all kinds of forecasting practice. To summarize, we can conclude the following: (1) in the individual forecasting model, the ANN model is the best in terms of RMSE (Table 5); (2) the combined forecasting models generally perform better than the individual forecasting models (Table 5); (3) the forecasting accuracy of nonlinear combination is better than that of linear combination (Table 5); (4) a combination of dissimilar methods outperforms a combination of similar methods (Table 6); (5) the PCA-based combination models perform better the other combination models and other individual models (Tables 5 7); and (6) the PANC model performs the best of all the linear (nonlinear) combination models and individual models (Tables 5 7). This leads to the final conclusion: (7) the novel nonlinear combination model can be used as an alternative tool for time series forecasting to obtain greater forecasting accuracy and improve the prediction quality further in view of empirical results. 5 Conclusions In this study, we propose a double-phase-processing procedure to treat the problem of whether to select or combine. Based on the proposed procedure, a nonlinear combination model is also proposed to overcome the drawbacks of existing combined methods. Simulation results obtained reveal that (1) the proposed procedure is useful, reliable and efficient for dealing with the problems of selecting or combining; (2) this proposed nonlinear combined model is also useful for improving the prediction performance of time series; and (3) thus, the proposed procedure and modeling technique can be a feasible solution for time series forecasting with multiple candidate models. However, several issues are generated and need to be addressed in the future. One important issue is the determination of the threshold values, such as the threshold value of performance stability and robustness metric, since this affects the decision-making judgments of researchers and practitioners. Another is the determination of the value of k in the computation of the robustness metric, the value of θ in the PCA process, and the size of the evaluation period in

16 16 YU LEAN et al. Vol. 18 the calculation of performance stability. In addition, the other reasonable judgmental rule of selecting one method or combining multiple methods, and more rational nonlinear combination technique, are worth exploring in further study of developments of the combination method. References [1] G. E. P. Box and G. Jenkins, Time Series Analysis: Forecasting and Control, Holden-Day, San Francisco, CA, [2] H. Akaike, Information theory and an extension of the maximum likelihood principle, In: The Proceedings of 2nd International Symposium of Information Theory (ed. by Petrov, B. N. and Csaki, F.), Akademia Kiado, Budapest, 1973, [3] G. Schwartz, Estimating the dimension of a model, The Annals of Statistics, 1978, 6: [4] L. Breiman, Stacked regressions, Machine Learning, 1996, 24: [5] Y. Yang, Combining time series models for forecasting, Working Paper, Department of Statistics, Iowa State University, [6] Y. Yang, Regression with multiple candidate models: Selecting or mixing? Working Paper, Department of Statistics, Iowa State University, [7] D. Draper, Assessment and propagation of model uncertainty, Journal of the Royal Statistical Society: Series B, 1995, 57: [8] E. I. George and R. E. McCulloch, Approaches for Bayesian variable selection, Statistica Sinica, 1997, 7: [9] A. E. Raftery, Bayesian model selection in social research (with discussion), In: Sociological Methodology (ed. by Marsden, P. V.), Blackwells, Cambridge, Massachusetts, 1995, [10] D. Madigan and J. York, Bayesian graphical models for discrete data, International Statistical Review, 1995, 63: [11] L. Breiman, Bagging predictors, Machine Learning, 1996, 24: [12] S. T. Buckland, K. P. Burnham and N. H. Augustin, Model selection: An integral part of inference, Biometrics, 1995, 53: [13] D. H. Wolpert, Stacked generalization, Neural Networks, 1992, 5: [14] M. LeBlanc and R. Tibshirani, Combining estimates in regression and classification, Journal of the American Statistical Association, 1996, 91: [15] A. Juditsky and A. Nemirovski, Functional aggregation for nonparametric estimation, The Annals of Statistics, 2000, 28 (3): [16] J. A. Hoeting, D. Madigan, A. E. Raftery and C. T. Volinsky, Bayesian model averaging: a tutorial, Statistical Science, 1999, 14(4): [17] N. Meade, Evidence for the selection of forecasting methods, Journal of Forecasting, 2000, 9: [18] A. Jessop, Sensitivity and robustness in selection problems, Computers and Operations Research, 2004, 31: [19] F. Diebold, Forecast combination and encompassing: reconciling two divergent literatures, International Journal of Forecasting, 1989, 5: [20] C. K. Chan, B. G. Kingsman and H. Wong, Determining when to update the weights in combined forecasts for product demand an application of the CUSUM technique, European Journal of Operational Research, 2004, 153: [21] M. Beccali, M. Cellura, V. Lo Brano and A. Marvuglia, Forecasting daily urban electric load profiles using artificial neural networks, Energy Conversion & Management, article in press, [22] N. Harvey and C. Harries, Effects of judges forecasting on their later combination of forecasts for the same outcomes, International Journal of Forecasting, article in press, [23] R. Lahdelma and H. Hakonen, An efficient linear programming algorithm for combined heat and power production, European Journal of Operational Research, 2003, 148: [24] A. J. Haklander and A. V. Delden, Thunderstorm predictors and their forecast skill for the Netherlands, Atmospheric Research, 2003, 67 68:

Selection of the Appropriate Lag Structure of Foreign Exchange Rates Forecasting Based on Autocorrelation Coefficient

Selection of the Appropriate Lag Structure of Foreign Exchange Rates Forecasting Based on Autocorrelation Coefficient Selection of the Appropriate Lag Structure of Foreign Exchange Rates Forecasting Based on Autocorrelation Coefficient Wei Huang 1,2, Shouyang Wang 2, Hui Zhang 3,4, and Renbin Xiao 1 1 School of Management,

More information

SARIMA-ELM Hybrid Model for Forecasting Tourist in Nepal

SARIMA-ELM Hybrid Model for Forecasting Tourist in Nepal Volume-03 Issue-07 July-2018 ISSN: 2455-3085 (Online) www.rrjournals.com [UGC Listed Journal] SARIMA-ELM Hybrid Model for Forecasting Tourist in Nepal *1 Kadek Jemmy Waciko & 2 Ismail B *1 Research Scholar,

More information

Combining time series models for forecasting

Combining time series models for forecasting International Journal of Forecasting 20 (2004) 69 84 www.elsevier.com/locate/ijforecast Combining time series models for forecasting Hui Zou a, Yuhong Yang b, * a Department of Statistics, Sequoia Hall,

More information

A Statistical Input Pruning Method for Artificial Neural Networks Used in Environmental Modelling

A Statistical Input Pruning Method for Artificial Neural Networks Used in Environmental Modelling A Statistical Input Pruning Method for Artificial Neural Networks Used in Environmental Modelling G. B. Kingston, H. R. Maier and M. F. Lambert Centre for Applied Modelling in Water Engineering, School

More information

Development of Stochastic Artificial Neural Networks for Hydrological Prediction

Development of Stochastic Artificial Neural Networks for Hydrological Prediction Development of Stochastic Artificial Neural Networks for Hydrological Prediction G. B. Kingston, M. F. Lambert and H. R. Maier Centre for Applied Modelling in Water Engineering, School of Civil and Environmental

More information

Empirical Approach to Modelling and Forecasting Inflation in Ghana

Empirical Approach to Modelling and Forecasting Inflation in Ghana Current Research Journal of Economic Theory 4(3): 83-87, 2012 ISSN: 2042-485X Maxwell Scientific Organization, 2012 Submitted: April 13, 2012 Accepted: May 06, 2012 Published: June 30, 2012 Empirical Approach

More information

22/04/2014. Economic Research

22/04/2014. Economic Research 22/04/2014 Economic Research Forecasting Models for Exchange Rate Tuesday, April 22, 2014 The science of prognostics has been going through a rapid and fruitful development in the past decades, with various

More information

Econometric Forecasting

Econometric Forecasting Graham Elliott Econometric Forecasting Course Description We will review the theory of econometric forecasting with a view to understanding current research and methods. By econometric forecasting we mean

More information

BAYESIAN ESTIMATION OF UNKNOWN PARAMETERS OVER NETWORKS

BAYESIAN ESTIMATION OF UNKNOWN PARAMETERS OVER NETWORKS BAYESIAN ESTIMATION OF UNKNOWN PARAMETERS OVER NETWORKS Petar M. Djurić Dept. of Electrical & Computer Engineering Stony Brook University Stony Brook, NY 11794, USA e-mail: petar.djuric@stonybrook.edu

More information

Do we need Experts for Time Series Forecasting?

Do we need Experts for Time Series Forecasting? Do we need Experts for Time Series Forecasting? Christiane Lemke and Bogdan Gabrys Bournemouth University - School of Design, Engineering and Computing Poole House, Talbot Campus, Poole, BH12 5BB - United

More information

Lossless Online Bayesian Bagging

Lossless Online Bayesian Bagging Lossless Online Bayesian Bagging Herbert K. H. Lee ISDS Duke University Box 90251 Durham, NC 27708 herbie@isds.duke.edu Merlise A. Clyde ISDS Duke University Box 90251 Durham, NC 27708 clyde@isds.duke.edu

More information

FORECASTING CHINA S FOREIGN TRADE VOLUME WITH A KERNEL-BASED HYBRID ECONOMETRIC-AI ENSEMBLE LEARNING APPROACH

FORECASTING CHINA S FOREIGN TRADE VOLUME WITH A KERNEL-BASED HYBRID ECONOMETRIC-AI ENSEMBLE LEARNING APPROACH Jrl Syst Sci & Complexity (2008) 21: 1 19 FORECASTING CHINA S FOREIGN TRADE VOLUME WITH A KERNEL-BASED HYBRID ECONOMETRIC-AI ENSEMBLE LEARNING APPROACH Lean YU Shouyang WANG Kin Keung LAI Received: 20

More information

interval forecasting

interval forecasting Interval Forecasting Based on Chapter 7 of the Time Series Forecasting by Chatfield Econometric Forecasting, January 2008 Outline 1 2 3 4 5 Terminology Interval Forecasts Density Forecast Fan Chart Most

More information

ARIMA modeling to forecast area and production of rice in West Bengal

ARIMA modeling to forecast area and production of rice in West Bengal Journal of Crop and Weed, 9(2):26-31(2013) ARIMA modeling to forecast area and production of rice in West Bengal R. BISWAS AND B. BHATTACHARYYA Department of Agricultural Statistics Bidhan Chandra Krishi

More information

A Modified Incremental Principal Component Analysis for On-line Learning of Feature Space and Classifier

A Modified Incremental Principal Component Analysis for On-line Learning of Feature Space and Classifier A Modified Incremental Principal Component Analysis for On-line Learning of Feature Space and Classifier Seiichi Ozawa, Shaoning Pang, and Nikola Kasabov Graduate School of Science and Technology, Kobe

More information

MODELLING TRAFFIC FLOW ON MOTORWAYS: A HYBRID MACROSCOPIC APPROACH

MODELLING TRAFFIC FLOW ON MOTORWAYS: A HYBRID MACROSCOPIC APPROACH Proceedings ITRN2013 5-6th September, FITZGERALD, MOUTARI, MARSHALL: Hybrid Aidan Fitzgerald MODELLING TRAFFIC FLOW ON MOTORWAYS: A HYBRID MACROSCOPIC APPROACH Centre for Statistical Science and Operational

More information

On Autoregressive Order Selection Criteria

On Autoregressive Order Selection Criteria On Autoregressive Order Selection Criteria Venus Khim-Sen Liew Faculty of Economics and Management, Universiti Putra Malaysia, 43400 UPM, Serdang, Malaysia This version: 1 March 2004. Abstract This study

More information

A NOTE ON THE HOMOMORPHISM THEOREM FOR HEMIRINGS

A NOTE ON THE HOMOMORPHISM THEOREM FOR HEMIRINGS Internal. J. Math. & Math. Sci. Vol. (1978)439-445 439 A NOTE ON THE HOMOMORPHISM THEOREM FOR HEMIRINGS D. M. OLSON Department of Mathematics Cameron University Lawton, Oklahoma 73501 U.S.A. (Recieved

More information

Testing for Regime Switching in Singaporean Business Cycles

Testing for Regime Switching in Singaporean Business Cycles Testing for Regime Switching in Singaporean Business Cycles Robert Breunig School of Economics Faculty of Economics and Commerce Australian National University and Alison Stegman Research School of Pacific

More information

A Comparison of Time Series Models for Forecasting Outbound Air Travel Demand *

A Comparison of Time Series Models for Forecasting Outbound Air Travel Demand * Journal of Aeronautics, Astronautics and Aviation, Series A, Vol.42, No.2 pp.073-078 (200) 73 A Comparison of Time Series Models for Forecasting Outbound Air Travel Demand * Yu-Wei Chang ** and Meng-Yuan

More information

Foreign Exchange Rates Forecasting with a C-Ascending Least Squares Support Vector Regression Model

Foreign Exchange Rates Forecasting with a C-Ascending Least Squares Support Vector Regression Model Foreign Exchange Rates Forecasting with a C-Ascending Least Squares Support Vector Regression Model Lean Yu, Xun Zhang, and Shouyang Wang Institute of Systems Science, Academy of Mathematics and Systems

More information

THE ROYAL STATISTICAL SOCIETY 2009 EXAMINATIONS SOLUTIONS GRADUATE DIPLOMA MODULAR FORMAT MODULE 3 STOCHASTIC PROCESSES AND TIME SERIES

THE ROYAL STATISTICAL SOCIETY 2009 EXAMINATIONS SOLUTIONS GRADUATE DIPLOMA MODULAR FORMAT MODULE 3 STOCHASTIC PROCESSES AND TIME SERIES THE ROYAL STATISTICAL SOCIETY 9 EXAMINATIONS SOLUTIONS GRADUATE DIPLOMA MODULAR FORMAT MODULE 3 STOCHASTIC PROCESSES AND TIME SERIES The Society provides these solutions to assist candidates preparing

More information

FORECASTING poor air quality events associated with

FORECASTING poor air quality events associated with A Comparison of Bayesian and Conditional Density Models in Probabilistic Ozone Forecasting Song Cai, William W. Hsieh, and Alex J. Cannon Member, INNS Abstract Probabilistic models were developed to provide

More information

A Modified Incremental Principal Component Analysis for On-Line Learning of Feature Space and Classifier

A Modified Incremental Principal Component Analysis for On-Line Learning of Feature Space and Classifier A Modified Incremental Principal Component Analysis for On-Line Learning of Feature Space and Classifier Seiichi Ozawa 1, Shaoning Pang 2, and Nikola Kasabov 2 1 Graduate School of Science and Technology,

More information

Bagging During Markov Chain Monte Carlo for Smoother Predictions

Bagging During Markov Chain Monte Carlo for Smoother Predictions Bagging During Markov Chain Monte Carlo for Smoother Predictions Herbert K. H. Lee University of California, Santa Cruz Abstract: Making good predictions from noisy data is a challenging problem. Methods

More information

DYNAMIC ECONOMETRIC MODELS Vol. 9 Nicolaus Copernicus University Toruń Mariola Piłatowska Nicolaus Copernicus University in Toruń

DYNAMIC ECONOMETRIC MODELS Vol. 9 Nicolaus Copernicus University Toruń Mariola Piłatowska Nicolaus Copernicus University in Toruń DYNAMIC ECONOMETRIC MODELS Vol. 9 Nicolaus Copernicus University Toruń 2009 Mariola Piłatowska Nicolaus Copernicus University in Toruń Combined Forecasts Using the Akaike Weights A b s t r a c t. The focus

More information

Univariate ARIMA Models

Univariate ARIMA Models Univariate ARIMA Models ARIMA Model Building Steps: Identification: Using graphs, statistics, ACFs and PACFs, transformations, etc. to achieve stationary and tentatively identify patterns and model components.

More information

DEPARTMENT OF ECONOMETRICS AND BUSINESS STATISTICS

DEPARTMENT OF ECONOMETRICS AND BUSINESS STATISTICS ISSN 1440-771X AUSTRALIA DEPARTMENT OF ECONOMETRICS AND BUSINESS STATISTICS Empirical Information Criteria for Time Series Forecasting Model Selection Md B Billah, R.J. Hyndman and A.B. Koehler Working

More information

Model Selection for Semiparametric Bayesian Models with Application to Overdispersion

Model Selection for Semiparametric Bayesian Models with Application to Overdispersion Proceedings 59th ISI World Statistics Congress, 25-30 August 2013, Hong Kong (Session CPS020) p.3863 Model Selection for Semiparametric Bayesian Models with Application to Overdispersion Jinfang Wang and

More information

Improved Holt Method for Irregular Time Series

Improved Holt Method for Irregular Time Series WDS'08 Proceedings of Contributed Papers, Part I, 62 67, 2008. ISBN 978-80-7378-065-4 MATFYZPRESS Improved Holt Method for Irregular Time Series T. Hanzák Charles University, Faculty of Mathematics and

More information

FORECASTING OF ECONOMIC QUANTITIES USING FUZZY AUTOREGRESSIVE MODEL AND FUZZY NEURAL NETWORK

FORECASTING OF ECONOMIC QUANTITIES USING FUZZY AUTOREGRESSIVE MODEL AND FUZZY NEURAL NETWORK FORECASTING OF ECONOMIC QUANTITIES USING FUZZY AUTOREGRESSIVE MODEL AND FUZZY NEURAL NETWORK Dusan Marcek Silesian University, Institute of Computer Science Opava Research Institute of the IT4Innovations

More information

le2 { Re d(a, f)l de > (1.2) n=2ar, 01 < e 2 ON ALPHA-CLOSE-TO-CONVEX FUNCTIONS OF ORDER BETA functions. f(z) z + z a n Re (1 ) zf + (I + zf"(z)) do >

le2 { Re d(a, f)l de > (1.2) n=2ar, 01 < e 2 ON ALPHA-CLOSE-TO-CONVEX FUNCTIONS OF ORDER BETA functions. f(z) z + z a n Re (1 ) zf + (I + zf(z)) do > Internat. J. Math. & Math. Sci. Vol. 9 No. 3 (1986) 435-438 435 ON ALPHA-CLOSE-TO-CONVEX FUNCTIONS OF ORDER BETA M.A. NASR Faculty of Science University of Mansoura Mansoura, Egypt (Received January 9,

More information

Econometric Forecasting

Econometric Forecasting Robert M. Kunst robert.kunst@univie.ac.at University of Vienna and Institute for Advanced Studies Vienna October 1, 2014 Outline Introduction Model-free extrapolation Univariate time-series models Trend

More information

Statistical Methods for Forecasting

Statistical Methods for Forecasting Statistical Methods for Forecasting BOVAS ABRAHAM University of Waterloo JOHANNES LEDOLTER University of Iowa John Wiley & Sons New York Chichester Brisbane Toronto Singapore Contents 1 INTRODUCTION AND

More information

Comparing the Univariate Modeling Techniques, Box-Jenkins and Artificial Neural Network (ANN) for Measuring of Climate Index

Comparing the Univariate Modeling Techniques, Box-Jenkins and Artificial Neural Network (ANN) for Measuring of Climate Index Applied Mathematical Sciences, Vol. 8, 2014, no. 32, 1557-1568 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2014.4150 Comparing the Univariate Modeling Techniques, Box-Jenkins and Artificial

More information

ECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications

ECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications ECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications Yongmiao Hong Department of Economics & Department of Statistical Sciences Cornell University Spring 2019 Time and uncertainty

More information

TIME SERIES ANALYSIS AND FORECASTING USING THE STATISTICAL MODEL ARIMA

TIME SERIES ANALYSIS AND FORECASTING USING THE STATISTICAL MODEL ARIMA CHAPTER 6 TIME SERIES ANALYSIS AND FORECASTING USING THE STATISTICAL MODEL ARIMA 6.1. Introduction A time series is a sequence of observations ordered in time. A basic assumption in the time series analysis

More information

Revisiting linear and non-linear methodologies for time series prediction - application to ESTSP 08 competition data

Revisiting linear and non-linear methodologies for time series prediction - application to ESTSP 08 competition data Revisiting linear and non-linear methodologies for time series - application to ESTSP 08 competition data Madalina Olteanu Universite Paris 1 - SAMOS CES 90 Rue de Tolbiac, 75013 Paris - France Abstract.

More information

Study on Modeling and Forecasting of the GDP of Manufacturing Industries in Bangladesh

Study on Modeling and Forecasting of the GDP of Manufacturing Industries in Bangladesh CHIANG MAI UNIVERSITY JOURNAL OF SOCIAL SCIENCE AND HUMANITIES M. N. A. Bhuiyan 1*, Kazi Saleh Ahmed 2 and Roushan Jahan 1 Study on Modeling and Forecasting of the GDP of Manufacturing Industries in Bangladesh

More information

Learning Static Parameters in Stochastic Processes

Learning Static Parameters in Stochastic Processes Learning Static Parameters in Stochastic Processes Bharath Ramsundar December 14, 2012 1 Introduction Consider a Markovian stochastic process X T evolving (perhaps nonlinearly) over time variable T. We

More information

Improved the Forecasting of ANN-ARIMA Model Performance: A Case Study of Water Quality at the Offshore Kuala Terengganu, Terengganu, Malaysia

Improved the Forecasting of ANN-ARIMA Model Performance: A Case Study of Water Quality at the Offshore Kuala Terengganu, Terengganu, Malaysia Improved the Forecasting of ANN-ARIMA Model Performance: A Case Study of Water Quality at the Offshore Kuala Terengganu, Terengganu, Malaysia Muhamad Safiih Lola1 Malaysia- safiihmd@umt.edu.my Mohd Noor

More information

Research Article Stacked Heterogeneous Neural Networks for Time Series Forecasting

Research Article Stacked Heterogeneous Neural Networks for Time Series Forecasting Hindawi Publishing Corporation Mathematical Problems in Engineering Volume 21, Article ID 373648, 2 pages doi:1.1155/21/373648 Research Article Stacked Heterogeneous Neural Networks for Time Series Forecasting

More information

FORECASTING SUGARCANE PRODUCTION IN INDIA WITH ARIMA MODEL

FORECASTING SUGARCANE PRODUCTION IN INDIA WITH ARIMA MODEL FORECASTING SUGARCANE PRODUCTION IN INDIA WITH ARIMA MODEL B. N. MANDAL Abstract: Yearly sugarcane production data for the period of - to - of India were analyzed by time-series methods. Autocorrelation

More information

Application of eigenvector-based spatial filtering approach to. a multinomial logit model for land use data

Application of eigenvector-based spatial filtering approach to. a multinomial logit model for land use data Presented at the Seventh World Conference of the Spatial Econometrics Association, the Key Bridge Marriott Hotel, Washington, D.C., USA, July 10 12, 2013. Application of eigenvector-based spatial filtering

More information

Introduction to Statistical modeling: handout for Math 489/583

Introduction to Statistical modeling: handout for Math 489/583 Introduction to Statistical modeling: handout for Math 489/583 Statistical modeling occurs when we are trying to model some data using statistical tools. From the start, we recognize that no model is perfect

More information

Electric Load Forecasting Using Wavelet Transform and Extreme Learning Machine

Electric Load Forecasting Using Wavelet Transform and Extreme Learning Machine Electric Load Forecasting Using Wavelet Transform and Extreme Learning Machine Song Li 1, Peng Wang 1 and Lalit Goel 1 1 School of Electrical and Electronic Engineering Nanyang Technological University

More information

Combining Forecasts: The End of the Beginning or the Beginning of the End? *

Combining Forecasts: The End of the Beginning or the Beginning of the End? * Published in International Journal of Forecasting (1989), 5, 585-588 Combining Forecasts: The End of the Beginning or the Beginning of the End? * Abstract J. Scott Armstrong The Wharton School, University

More information

AlmaTourism N. 2, 2010 : Air Passenger Flows: Evidence from Sicily and Sardinia. Contents liste available at Cib.Unibo.

AlmaTourism N. 2, 2010 : Air Passenger Flows: Evidence from Sicily and Sardinia. Contents liste available at Cib.Unibo. Contents liste available at Cib.Unibo AlmaTourism homepage: almatourism.cib.unibo.it Air Passenger Flows: Evidence from Sicily and Sardinia Castellani, M. Department of Economics, University of Bologna

More information

Automatic forecasting with a modified exponential smoothing state space framework

Automatic forecasting with a modified exponential smoothing state space framework ISSN 1440-771X Department of Econometrics and Business Statistics http://www.buseco.monash.edu.au/depts/ebs/pubs/wpapers/ Automatic forecasting with a modified exponential smoothing state space framework

More information

Forecasting Bangladesh's Inflation through Econometric Models

Forecasting Bangladesh's Inflation through Econometric Models American Journal of Economics and Business Administration Original Research Paper Forecasting Bangladesh's Inflation through Econometric Models 1,2 Nazmul Islam 1 Department of Humanities, Bangladesh University

More information

Comparing Robustness of Pairwise and Multiclass Neural-Network Systems for Face Recognition

Comparing Robustness of Pairwise and Multiclass Neural-Network Systems for Face Recognition Comparing Robustness of Pairwise and Multiclass Neural-Network Systems for Face Recognition J. Uglov, V. Schetinin, C. Maple Computing and Information System Department, University of Bedfordshire, Luton,

More information

Performance of Autoregressive Order Selection Criteria: A Simulation Study

Performance of Autoregressive Order Selection Criteria: A Simulation Study Pertanika J. Sci. & Technol. 6 (2): 7-76 (2008) ISSN: 028-7680 Universiti Putra Malaysia Press Performance of Autoregressive Order Selection Criteria: A Simulation Study Venus Khim-Sen Liew, Mahendran

More information

Short-term water demand forecast based on deep neural network ABSTRACT

Short-term water demand forecast based on deep neural network ABSTRACT Short-term water demand forecast based on deep neural network Guancheng Guo 1, Shuming Liu 2 1,2 School of Environment, Tsinghua University, 100084, Beijing, China 2 shumingliu@tsinghua.edu.cn ABSTRACT

More information

Modified Holt s Linear Trend Method

Modified Holt s Linear Trend Method Modified Holt s Linear Trend Method Guckan Yapar, Sedat Capar, Hanife Taylan Selamlar and Idil Yavuz Abstract Exponential smoothing models are simple, accurate and robust forecasting models and because

More information

A SEASONAL FUZZY TIME SERIES FORECASTING METHOD BASED ON GUSTAFSON-KESSEL FUZZY CLUSTERING *

A SEASONAL FUZZY TIME SERIES FORECASTING METHOD BASED ON GUSTAFSON-KESSEL FUZZY CLUSTERING * No.2, Vol.1, Winter 2012 2012 Published by JSES. A SEASONAL FUZZY TIME SERIES FORECASTING METHOD BASED ON GUSTAFSON-KESSEL * Faruk ALPASLAN a, Ozge CAGCAG b Abstract Fuzzy time series forecasting methods

More information

FORECASTING TIME SERIES DATA USING HYBRID GREY RELATIONAL ARTIFICIAL NEURAL NETWORK AND AUTO REGRESSIVE INTEGRATED MOVING AVERAGE MODEL

FORECASTING TIME SERIES DATA USING HYBRID GREY RELATIONAL ARTIFICIAL NEURAL NETWORK AND AUTO REGRESSIVE INTEGRATED MOVING AVERAGE MODEL FORECASTING TIME SERIES DATA USING HYBRID GREY RELATIONAL ARTIFICIAL NEURAL NETWORK AND AUTO REGRESSIVE INTEGRATED MOVING AVERAGE MODEL Roselina Sallehuddin, Siti Mariyam Hj. Shamsuddin, Siti Zaiton Mohd.

More information

Learning Gaussian Process Models from Uncertain Data

Learning Gaussian Process Models from Uncertain Data Learning Gaussian Process Models from Uncertain Data Patrick Dallaire, Camille Besse, and Brahim Chaib-draa DAMAS Laboratory, Computer Science & Software Engineering Department, Laval University, Canada

More information

at least 50 and preferably 100 observations should be available to build a proper model

at least 50 and preferably 100 observations should be available to build a proper model III Box-Jenkins Methods 1. Pros and Cons of ARIMA Forecasting a) need for data at least 50 and preferably 100 observations should be available to build a proper model used most frequently for hourly or

More information

Stochastic Hydrology. a) Data Mining for Evolution of Association Rules for Droughts and Floods in India using Climate Inputs

Stochastic Hydrology. a) Data Mining for Evolution of Association Rules for Droughts and Floods in India using Climate Inputs Stochastic Hydrology a) Data Mining for Evolution of Association Rules for Droughts and Floods in India using Climate Inputs An accurate prediction of extreme rainfall events can significantly aid in policy

More information

A hybrid model of stock price prediction based on the PCA-ARIMA-BP

A hybrid model of stock price prediction based on the PCA-ARIMA-BP ANZIAM J. 58 (E) pp.e162 E178, 2017 E162 A hybrid model of stock price prediction based on the PCA-ARIMA-BP Hua Luo 1 Shuang Wang 2 (Received 13 June 2016; revised 17 February 2017) Abstract A pca-arima-bp

More information

Seasonal Autoregressive Integrated Moving Average Model for Precipitation Time Series

Seasonal Autoregressive Integrated Moving Average Model for Precipitation Time Series Journal of Mathematics and Statistics 8 (4): 500-505, 2012 ISSN 1549-3644 2012 doi:10.3844/jmssp.2012.500.505 Published Online 8 (4) 2012 (http://www.thescipub.com/jmss.toc) Seasonal Autoregressive Integrated

More information

8.6 Bayesian neural networks (BNN) [Book, Sect. 6.7]

8.6 Bayesian neural networks (BNN) [Book, Sect. 6.7] 8.6 Bayesian neural networks (BNN) [Book, Sect. 6.7] While cross-validation allows one to find the weight penalty parameters which would give the model good generalization capability, the separation of

More information

A Comparison of the Forecast Performance of. Double Seasonal ARIMA and Double Seasonal. ARFIMA Models of Electricity Load Demand

A Comparison of the Forecast Performance of. Double Seasonal ARIMA and Double Seasonal. ARFIMA Models of Electricity Load Demand Applied Mathematical Sciences, Vol. 6, 0, no. 35, 6705-67 A Comparison of the Forecast Performance of Double Seasonal ARIMA and Double Seasonal ARFIMA Models of Electricity Load Demand Siti Normah Hassan

More information

Performance of Cross Validation in Tree-Based Models

Performance of Cross Validation in Tree-Based Models Performance of Cross Validation in Tree-Based Models Seoung Bum Kim, Xiaoming Huo, Kwok-Leung Tsui School of Industrial and Systems Engineering Georgia Institute of Technology Atlanta, Georgia 30332 {sbkim,xiaoming,ktsui}@isye.gatech.edu

More information

Structural Uncertainty in Health Economic Decision Models

Structural Uncertainty in Health Economic Decision Models Structural Uncertainty in Health Economic Decision Models Mark Strong 1, Hazel Pilgrim 1, Jeremy Oakley 2, Jim Chilcott 1 December 2009 1. School of Health and Related Research, University of Sheffield,

More information

Click Prediction and Preference Ranking of RSS Feeds

Click Prediction and Preference Ranking of RSS Feeds Click Prediction and Preference Ranking of RSS Feeds 1 Introduction December 11, 2009 Steven Wu RSS (Really Simple Syndication) is a family of data formats used to publish frequently updated works. RSS

More information

On Mixture Regression Shrinkage and Selection via the MR-LASSO

On Mixture Regression Shrinkage and Selection via the MR-LASSO On Mixture Regression Shrinage and Selection via the MR-LASSO Ronghua Luo, Hansheng Wang, and Chih-Ling Tsai Guanghua School of Management, Peing University & Graduate School of Management, University

More information

Least Absolute Shrinkage is Equivalent to Quadratic Penalization

Least Absolute Shrinkage is Equivalent to Quadratic Penalization Least Absolute Shrinkage is Equivalent to Quadratic Penalization Yves Grandvalet Heudiasyc, UMR CNRS 6599, Université de Technologie de Compiègne, BP 20.529, 60205 Compiègne Cedex, France Yves.Grandvalet@hds.utc.fr

More information

High Wind and Energy Specific Models for Global. Production Forecast

High Wind and Energy Specific Models for Global. Production Forecast High Wind and Energy Specific Models for Global Production Forecast Carlos Alaíz, Álvaro Barbero, Ángela Fernández, José R. Dorronsoro Dpto. de Ingeniería Informática and Instituto de Ingeniería del Conocimiento

More information

Test for Parameter Change in ARIMA Models

Test for Parameter Change in ARIMA Models Test for Parameter Change in ARIMA Models Sangyeol Lee 1 Siyun Park 2 Koichi Maekawa 3 and Ken-ichi Kawai 4 Abstract In this paper we consider the problem of testing for parameter changes in ARIMA models

More information

GDP forecast errors Satish Ranchhod

GDP forecast errors Satish Ranchhod GDP forecast errors Satish Ranchhod Editor s note This paper looks more closely at our forecasts of growth in Gross Domestic Product (GDP). It considers two different measures of GDP, production and expenditure,

More information

Chapter 3: Regression Methods for Trends

Chapter 3: Regression Methods for Trends Chapter 3: Regression Methods for Trends Time series exhibiting trends over time have a mean function that is some simple function (not necessarily constant) of time. The example random walk graph from

More information

Master of Science in Statistics A Proposal

Master of Science in Statistics A Proposal 1 Master of Science in Statistics A Proposal Rationale of the Program In order to cope up with the emerging complexity on the solutions of realistic problems involving several phenomena of nature it is

More information

Warwick Business School Forecasting System. Summary. Ana Galvao, Anthony Garratt and James Mitchell November, 2014

Warwick Business School Forecasting System. Summary. Ana Galvao, Anthony Garratt and James Mitchell November, 2014 Warwick Business School Forecasting System Summary Ana Galvao, Anthony Garratt and James Mitchell November, 21 The main objective of the Warwick Business School Forecasting System is to provide competitive

More information

MS-C1620 Statistical inference

MS-C1620 Statistical inference MS-C1620 Statistical inference 10 Linear regression III Joni Virta Department of Mathematics and Systems Analysis School of Science Aalto University Academic year 2018 2019 Period III - IV 1 / 32 Contents

More information

Using Temporal Hierarchies to Predict Tourism Demand

Using Temporal Hierarchies to Predict Tourism Demand Using Temporal Hierarchies to Predict Tourism Demand International Symposium on Forecasting 29 th June 2015 Nikolaos Kourentzes Lancaster University George Athanasopoulos Monash University n.kourentzes@lancaster.ac.uk

More information

That s Hot: Predicting Daily Temperature for Different Locations

That s Hot: Predicting Daily Temperature for Different Locations That s Hot: Predicting Daily Temperature for Different Locations Alborz Bejnood, Max Chang, Edward Zhu Stanford University Computer Science 229: Machine Learning December 14, 2012 1 Abstract. The problem

More information

The Lady Tasting Tea. How to deal with multiple testing. Need to explore many models. More Predictive Modeling

The Lady Tasting Tea. How to deal with multiple testing. Need to explore many models. More Predictive Modeling The Lady Tasting Tea More Predictive Modeling R. A. Fisher & the Lady B. Muriel Bristol claimed she prefers tea added to milk rather than milk added to tea Fisher was skeptical that she could distinguish

More information

Selecting an optimal set of parameters using an Akaike like criterion

Selecting an optimal set of parameters using an Akaike like criterion Selecting an optimal set of parameters using an Akaike like criterion R. Moddemeijer a a University of Groningen, Department of Computing Science, P.O. Box 800, L-9700 AV Groningen, The etherlands, e-mail:

More information

A Non-Parametric Approach of Heteroskedasticity Robust Estimation of Vector-Autoregressive (VAR) Models

A Non-Parametric Approach of Heteroskedasticity Robust Estimation of Vector-Autoregressive (VAR) Models Journal of Finance and Investment Analysis, vol.1, no.1, 2012, 55-67 ISSN: 2241-0988 (print version), 2241-0996 (online) International Scientific Press, 2012 A Non-Parametric Approach of Heteroskedasticity

More information

A Hybrid Time-delay Prediction Method for Networked Control System

A Hybrid Time-delay Prediction Method for Networked Control System International Journal of Automation and Computing 11(1), February 2014, 19-24 DOI: 10.1007/s11633-014-0761-1 A Hybrid Time-delay Prediction Method for Networked Control System Zhong-Da Tian Xian-Wen Gao

More information

FORECASTING COARSE RICE PRICES IN BANGLADESH

FORECASTING COARSE RICE PRICES IN BANGLADESH Progress. Agric. 22(1 & 2): 193 201, 2011 ISSN 1017-8139 FORECASTING COARSE RICE PRICES IN BANGLADESH M. F. Hassan*, M. A. Islam 1, M. F. Imam 2 and S. M. Sayem 3 Department of Agricultural Statistics,

More information

Autoregressive Integrated Moving Average Model to Predict Graduate Unemployment in Indonesia

Autoregressive Integrated Moving Average Model to Predict Graduate Unemployment in Indonesia DOI 10.1515/ptse-2017-0005 PTSE 12 (1): 43-50 Autoregressive Integrated Moving Average Model to Predict Graduate Unemployment in Indonesia Umi MAHMUDAH u_mudah@yahoo.com (State Islamic University of Pekalongan,

More information

Bootstrap for model selection: linear approximation of the optimism

Bootstrap for model selection: linear approximation of the optimism Bootstrap for model selection: linear approximation of the optimism G. Simon 1, A. Lendasse 2, M. Verleysen 1, Université catholique de Louvain 1 DICE - Place du Levant 3, B-1348 Louvain-la-Neuve, Belgium,

More information

Lesson 13: Box-Jenkins Modeling Strategy for building ARMA models

Lesson 13: Box-Jenkins Modeling Strategy for building ARMA models Lesson 13: Box-Jenkins Modeling Strategy for building ARMA models Facoltà di Economia Università dell Aquila umberto.triacca@gmail.com Introduction In this lesson we present a method to construct an ARMA(p,

More information

Pattern Matching and Neural Networks based Hybrid Forecasting System

Pattern Matching and Neural Networks based Hybrid Forecasting System Pattern Matching and Neural Networks based Hybrid Forecasting System Sameer Singh and Jonathan Fieldsend PA Research, Department of Computer Science, University of Exeter, Exeter, UK Abstract In this paper

More information

Least Squares Model Averaging. Bruce E. Hansen University of Wisconsin. January 2006 Revised: August 2006

Least Squares Model Averaging. Bruce E. Hansen University of Wisconsin. January 2006 Revised: August 2006 Least Squares Model Averaging Bruce E. Hansen University of Wisconsin January 2006 Revised: August 2006 Introduction This paper developes a model averaging estimator for linear regression. Model averaging

More information

Evaluation of Some Techniques for Forecasting of Electricity Demand in Sri Lanka

Evaluation of Some Techniques for Forecasting of Electricity Demand in Sri Lanka Appeared in Sri Lankan Journal of Applied Statistics (Volume 3) 00 Evaluation of Some echniques for Forecasting of Electricity Demand in Sri Lanka.M.J. A. Cooray and M.Indralingam Department of Mathematics

More information

Consistency of test based method for selection of variables in high dimensional two group discriminant analysis

Consistency of test based method for selection of variables in high dimensional two group discriminant analysis https://doi.org/10.1007/s42081-019-00032-4 ORIGINAL PAPER Consistency of test based method for selection of variables in high dimensional two group discriminant analysis Yasunori Fujikoshi 1 Tetsuro Sakurai

More information

Structure learning in human causal induction

Structure learning in human causal induction Structure learning in human causal induction Joshua B. Tenenbaum & Thomas L. Griffiths Department of Psychology Stanford University, Stanford, CA 94305 jbt,gruffydd @psych.stanford.edu Abstract We use

More information

Analysis Methods for Supersaturated Design: Some Comparisons

Analysis Methods for Supersaturated Design: Some Comparisons Journal of Data Science 1(2003), 249-260 Analysis Methods for Supersaturated Design: Some Comparisons Runze Li 1 and Dennis K. J. Lin 2 The Pennsylvania State University Abstract: Supersaturated designs

More information

1.3 STATISTICAL WIND POWER FORECASTING FOR U.S. WIND FARMS

1.3 STATISTICAL WIND POWER FORECASTING FOR U.S. WIND FARMS 1.3 STATISTICAL WIND POWER FORECASTING FOR U.S. WIND FARMS Michael Milligan, Consultant * Marc Schwartz and Yih-Huei Wan National Renewable Energy Laboratory, Golden, Colorado ABSTRACT Electricity markets

More information

Model Averaging With Holdout Estimation of the Posterior Distribution

Model Averaging With Holdout Estimation of the Posterior Distribution Model Averaging With Holdout stimation of the Posterior Distribution Alexandre Lacoste alexandre.lacoste.1@ulaval.ca François Laviolette francois.laviolette@ift.ulaval.ca Mario Marchand mario.marchand@ift.ulaval.ca

More information

Forecasting Major Vegetable Crops Productions in Tunisia

Forecasting Major Vegetable Crops Productions in Tunisia International Journal of Research in Business Studies and Management Volume 2, Issue 6, June 2015, PP 15-19 ISSN 2394-5923 (Print) & ISSN 2394-5931 (Online) Forecasting Major Vegetable Crops Productions

More information

Finding Robust Solutions to Dynamic Optimization Problems

Finding Robust Solutions to Dynamic Optimization Problems Finding Robust Solutions to Dynamic Optimization Problems Haobo Fu 1, Bernhard Sendhoff, Ke Tang 3, and Xin Yao 1 1 CERCIA, School of Computer Science, University of Birmingham, UK Honda Research Institute

More information

FORECASTING THE INVENTORY LEVEL OF MAGNETIC CARDS IN TOLLING SYSTEM

FORECASTING THE INVENTORY LEVEL OF MAGNETIC CARDS IN TOLLING SYSTEM FORECASTING THE INVENTORY LEVEL OF MAGNETIC CARDS IN TOLLING SYSTEM Bratislav Lazić a, Nebojša Bojović b, Gordana Radivojević b*, Gorana Šormaz a a University of Belgrade, Mihajlo Pupin Institute, Serbia

More information

Automatic Forecasting

Automatic Forecasting Automatic Forecasting Summary The Automatic Forecasting procedure is designed to forecast future values of time series data. A time series consists of a set of sequential numeric data taken at equally

More information

Stochastic Analogues to Deterministic Optimizers

Stochastic Analogues to Deterministic Optimizers Stochastic Analogues to Deterministic Optimizers ISMP 2018 Bordeaux, France Vivak Patel Presented by: Mihai Anitescu July 6, 2018 1 Apology I apologize for not being here to give this talk myself. I injured

More information

Deep Learning Architecture for Univariate Time Series Forecasting

Deep Learning Architecture for Univariate Time Series Forecasting CS229,Technical Report, 2014 Deep Learning Architecture for Univariate Time Series Forecasting Dmitry Vengertsev 1 Abstract This paper studies the problem of applying machine learning with deep architecture

More information

Application of Time Sequence Model Based on Excluded Seasonality in Daily Runoff Prediction

Application of Time Sequence Model Based on Excluded Seasonality in Daily Runoff Prediction Send Orders for Reprints to reprints@benthamscience.ae 546 The Open Cybernetics & Systemics Journal, 2014, 8, 546-552 Open Access Application of Time Sequence Model Based on Excluded Seasonality in Daily

More information