Foreign Exchange Rates Forecasting with a C-Ascending Least Squares Support Vector Regression Model
|
|
- Dortha Cox
- 5 years ago
- Views:
Transcription
1 Foreign Exchange Rates Forecasting with a C-Ascending Least Squares Support Vector Regression Model Lean Yu, Xun Zhang, and Shouyang Wang Institute of Systems Science, Academy of Mathematics and Systems Science, Chinese Academy of Sciences, Beijing China {yulean,zhangxun,sywang}@amss.ac.cn Abstract. In this paper, a modified least squares support vector regression (LSSVR) model, called C-ascending least squares support vector regression (C-ALSSVR), is proposed for foreign exchange rates forecasting. The generic idea of the proposed C-ALSSVR model is based on the prior knowledge that different data points often provide different information for modeling and more weights should be given to those data points containing more information. The C-ALSSVR can be obtained by a simple modification of the regularization parameter in LSSVR, whereby more weights are given to the recent least squares errors than the distant least squares errors while keeping the regularized terms in its original form. For verification purpose, the performance of the C-ALSSVR model is evaluated using three typical foreign exchange rates. Experimental results obtained demonstrated that the C-ALSSVR model is very promising tool in foreign exchange rates forecasting. Keywords: Foreign exchange rates forecasting, least squares support vector regression, regularization parameter. 1 Introduction The foreign exchange market is a nonlinear dynamic market with high volatility and thus foreign exchange rate series are inherently noisy, non-stationary and deterministically chaotic [1]. Due to its irregularity, foreign exchange rates forecasting is regarded as a rather challenging task. For traditional linear-based forecasting methods such as autoregressive integrated moving average (ARIMA) and exponential smoothing model (ESM) [2], it is extremely difficult to capture the irregularity because they are unable to capture subtle nonlinear patterns hidden in the foreign exchange rate series data. To remedy the gap, many emerging nonlinear techniques, such as artificial neural networks (ANNs), were widely used in the foreign exchange rates forecasting and obtained good results relative to the traditional linear modeling techniques in the past decades. For example, De Matos [3] compared the strength of a multilayer feed-forward neural network (MLFNN) with that of a recurrent neural G. Allen et al. (Eds.): ICCS 2009, Part II, LNCS 5545, pp , c Springer-Verlag Berlin Heidelberg 2009
2 Foreign Exchange Rates Forecasting with a C-ALSSVR Model 607 network (RNN) based on the forecasting of Japanese yen futures. Kuan and Liu [4] provided a comparative evaluation of the performance of MLFNN and a RNN on the prediction of an array of commonly traded exchange rates. In the article of Tenti [5], the RNN is directly applied to exchange rates forecasting. Hsu et al. [6] developed a clustering neural network (CNN) model to predict the direction of movements in the USD/DEM exchange rates. Their experimental results suggested that their proposed model achieved better forecasting performance relative to other indicators. In a more recent study by Leung et al. [7], the forecasting accuracy of MLFNN was compared with the general regression neural network (GRNN). The study showed that the GRNN possessed a greater forecasting strength relative to MLFNN with respect to a variety of currency exchange rates. Similarly, Chen and Leung [8] adopted an error correction neural network (ECNN) model to predict foreign exchange rates. Yu et al. [9] proposed an adaptive smoothing neural network (ASNN) model by adaptively adjusting error signals to predict foreign exchange rates and obtained good performance. Although the ANN models achieve great success in foreign exchange rates forecasting, they still have some disadvantages in some practical applications. For example, ANN models often suffer from over-fitting problem in the case of when training was performed too long or where training examples are rare. Furthermore, local minimum problem often occurred due to the adoption of empirical risk minimization (ERM) principle in ANN learning. For this purpose, a competitive neural network learning model, called support vector machines (SVMs), was proposed by Vapnik and his colleagues in 1995 [10]. The SVM is a novel learning way to train polynomial neural networks based on the structural risk minimization (SRM) principle where seeks to minimize an upper bound of the generalization error rather than minimize the empirical error implemented in other neural networks. The generic idea is based on the fact that the generalization error is bounded by the sum of the empirical error and a confidence interval term that depends on the Vapnik-Chervonenkis (VC) dimension [11]. Using the SRM principle, the SVM will obtain a global optimum solution by adopting a suitable trade-off between the empirical error and the VC-confidence interval. Due to this distinct characteristic, the SVM has been widely applied to pattern classification and function approximation or regression estimation problems [12]. In terms of the classification and regression problems, the SVM can be categorized into support vector classification (SVC) and support vector regression (SVR). In this paper, we focus the SVR for foreign exchange rates forecasting. In SVR, the problem is formulated by employing the so-called Vapnik s ε-insensitive loss function, taking the regression problem as an inequality constrained convex programming problem, more specifically a quadratic programming (QP) problem and using the Mercer condition for mapping from nonlinear feature space to the chosen kernel function [13]. Usually, this QP can lead to higher computational cost. For this purpose, least squares support vector machine (LSSVM), as a variant of SVM, tries to avoid the above shortcoming and obtain an analytical solution directly from solving a set of linear equations
3 608 L. Yu, X. Zhang, and S. Wang instead of solving QP problem. In such a way, a least squares support vector regression (LSSVR) can be formulated by replacing Vapnik s ε-insensitive loss function with a least squares cost function corresponding to a form of ridge regression [13]. By introducing the least squares cost function, the LSSVM can be extended to solve nonlinear regression problems. In the practical time series forecasting, many studies have shown that the relationship between input variables and output variable gradually changed over the time and recent data often contain more information than distant data. It is therefore advisable for us to give more weights to the recent containing more information. In view of this idea, an innovative approach is proposed by Tay and Cao [11] which used the ascending regularization parameter in the SVR to predict financial time series based on the work of Refenes and Bentz [14]. The ascending SVR is obtained by a simple modification of the regularization risk function in SVR, whereby the recent ε-insensitive errors are penalized more heavily than the distant ε-insensitive errors. The ascending SVR is reported to be very effective in financial time series forecasting. This paper is motivated by the ascending SVR model, and generalizes the idea for LSSVR whereby more weights are given to the recent least squares errors than the distant lease squares errors in the regularization parameter. The primary objective of this paper is to propose a new forecasting paradigm called C-ascending LSSVR that can significantly reduce the computational cost and to improve the prediction capability of standard SVR as well as to examine whether the prior knowledge that recent data should provide more information than distant data can also be utilized by LSSVR in financial time series forecasting, especially for foreign exchange rates forecasting in this study. The remainder of this paper is organized as follows. Section 2 overviews the formulation of least square support vector regression (LSSVR) model briefly. In Section 3, the formulation of the C-ascending LSSVR (C-ALSSVR) is presented in detail. For further illustration, three typical foreign exchange rates prediction experiments are conducted in Section 4. Finally, some concluding remarks are drawninsection5. 2 Least Squares Support Vector Regression (LSSVR) Suppose that there is a give training set D of n data points {(x i,y i )} n i=1 with input data x i x R n and output y i y R. An important idea of SVM is to map the input into a high-dimensional feature space F via a nonlinear mapping function ϕ(x) to find an unknown function form f, which takes the following form y = f(x; w, b) =w T ϕ(x)+b (1) where ϕ( ) is a nonlinear mapping function, w i is the ith weights, and b is a bias. In the least squares support vector regression, the following optimization problem is formulated min J = 1 2 wt w + C 2 n i=1 e2 i (2)
4 Foreign Exchange Rates Forecasting with a C-ALSSVR Model 609 Subject to the following equality constraints y i = w T ϕ(x i )+b + e i, for i =1, 2,,n. (3) where C is the regularization parameter, e i is the ith approximation error between predicted and actual values. Combining (2) and (3), one can define a Lagrangian function L(w, b, e i ; α i )= 1 2 wt w + C 2 n i=1 e2 i n i=1 α i [ w T ϕ(x i )+b + e i y i ] (4) where α i is the ith Lagrangian multiplier. The optimal conditions are obtained by differentiating (4) L w =0 w = n α iϕ(x i ) i=1 L b =0 n α i =0 i=1 L (5) e =0 e i = α i /C L =0 w T ϕ(x i )+b + e i y i =0 α i for i =1, 2,,n. After elimination of e i and w, the solution is given by the following set of linear equations { n i,j=1 α iϕ(x i ) T ϕ(x j )+b +(α i /C) y i =0 n i=1 α (6) i =0 for i, j =1, 2,,n. Using the Mercer condition, the kernel function can be defined as K(x i,x j )=ϕ(x i ) T ϕ(x j )fori, j =1, 2,,n.Typical kernel functions include linear kernel K(x i,x j )=x T i x j, polynomial kernel K(x i,x j )= ( x T i x j +1 ) d, Gaussian kernel or RBF kernel K(x i,x j )=exp ( x i x j 2/ ) σ 2,andMLP kernel K(x i,x j )=tanh ( kx T i x j + θ ) where d, σ, k and θ are kernel parameters, which are specified by users beforehand. Accordingly, (6) can be rewritten as { Ωα + b1=y T (7) 1 α =0 where b is a scalar, Ω, α, y, and 1 are either matrix or vectors, which are defined, respectively, Ω = K(x i,x j )+(1/C)I, (8) α =(α 1,α 2,,α n ) T, (9) y =(y 1,y 2,,y n ) T, (10) 1=(1, 1,, 1) T (11)
5 610 L. Yu, X. Zhang, and S. Wang where I is a unit matrix in (8). Equivalently, using the matrix form, the linear equations of (6) can be expressed by [ ] [ ] [ ] Ω 1 α y T = (12) 1 0 b 0 From (8), the Ω is positive definite, the solution of Lagrangian multiplier α can be obtained from (12), i.e., α = Ω 1 (y b 1) (13) Substituting (13) into the second matrix equation in (12), we can obtain b = 1 T Ω 1 y 1 T Ω 1 1 (14) Here, since Ω is positive definite, Ω 1 is also positive definite and thus 1 T Ω 1 1 > 0. Thus, b is always obtained. Substituting (14) into (13), α can be easily obtained. Accordingly the solution of w can be obtained from the first equation in (5). Using w and b, the function f shownin(1)canbedetermined. The distinct advantages of LSSVR reflect the following two-fold. On the one hand, the optimal solution of (1) can be found by solving a set of linear equations instead of solving a quadratic programming (QP) problem which is used in standard SVR thus reducing the computational costs. On the other hand, relative to the standard SVR with the Vapnik s ε-insensitive loss function, there is no need to determine an additional ε accuracy parameter which is related to ε-insensitive loss function and reducing the chance of over-fitting. 3 C-Ascending Least Squares Support Vector Regression As is known to many researchers in the field of machine learning, the regularization parameter C shown in (2) determines the trade-off between the regularized term and the tolerable empirical errors. With the increase of C, the relative importance of empirical errors will grow relative to the regularized term, and vice versa. Usually, in standard SVR and LSSVR, the empirical risk function has equal weight C to all ε-insensitive loss function [11] and least squares errors e i (i =1,2,...,n) between the predicted and actual values [13]. That is, the regularization parameter C is a constant or a fixed value. However, many time series forecasting experiments (e.g., [11]) have shown that a fixed regularization parameter is unsuitable for some prediction tasks with some prior knowledge. Considering that different data might contain different information, more weights should be given to those data offering more information. In the case of LSSVR for financial time series forecasting, more weights should be given to the recent data taking the prior knowledge into account that recent data might offer more information than distant data. For this
6 Foreign Exchange Rates Forecasting with a C-ALSSVR Model 611 purpose, the regularization parameter C should be replaced by a variable regularization parameter C i to capture the variation of information containing in thetimeseriesdata{(x i,y i )} n i=1. In terms of the prior knowledge that recent data might offer more information than distant data, the variable regularization parameter C i should satisfy C i >C i 1 (i = 2,..., n). Since the variable regularization parameters C i will grow from the distant data points to the recent data points, C i is called ascending regularization parameter which will give more weights on the more recent data points. In the practical applications, the form of C i often depends on the prior knowledge we have. In the financial time series forecasting, two typical forms: linear form and exponential form [11] are often used. For the linear ascending form, the number of training data points are often used to determine the value of C i. A common linear form is shown as follows: i C i = n(n +1)/2 C = 2i n(n +1) C (15) where i is a time factor, n is the number of training data point, and C is a constant that needs tuning. Usually the more recent the training data point is, the larger the regularization parameter C i is, according to (15). A distinct advantage of linear form is simplicity and easy to implementation. When the size of training data set is small and low computational cost is expected, it is a good choice. For the exponential ascending form, another parameter r is introduced into the exponential function in order to control the rate of ascending, which is represented as i C i = 1+exp(r 2ri/n) C (16) Exponential form could offer more ways to describe the changing of the information containing in the training data. When the size of training data set is large and the computational cost is not concerned, the exponential form can be adopted. Besides the linear and exponential forms, other forms that can control the ascending rates can also be employed in the practical application. Based on the ascending regularization parameter C i, a new LSSVR called C- ascending LSSVR (C-ALSSVR) can be introduced. Similar to (2) and (3), the optimization problem of C-ALSSVR for time series prediction can be formulated as follows. { min J = 1 2 wt w + Ci n 2 i=1 e2 i s.t. y i = w T (17) ϕ(x i )+b + e i,fori =1, 2,,n. Using the Lagrangian theorem, the final solution is similar to (13) and (14). The only difference is the value of Ω due to the introduction of the variable regularization parameter C i.inthecaseofc-alssvr, the value of Ω is calculated by Ω = K(x i,x j )+(1/C i )I (18) According to (17) and (18), the LSSVR algorithm can still be used except the regularization parameter value C i for every training data points is different, and
7 612 L. Yu, X. Zhang, and S. Wang thus computation and simulation procedures of LSSVR should be utilized by a simple modification of regularization parameter from a fixed value C to a variable parameter C i in terms of (17). 4 Experimental Results In this section, three real-world foreign exchange rates are used to test the effectiveness of the proposed C-ascending LSSVR model. The data used here are monthly and are obtained from Pacific Exchange Rates Services, provided by Professor Werner Antweiler, University of British Columbia, Vancouver, Canada. They consist of the US dollar against each of the three currencies British pounds (GBP), euros (EUR) and Japanese yen (JPY) studied in this paper. We take monthly data from January 1971 to December 2000 as in-sample (training periods) data sets (360 observations including 60 samples for cross-validations). We also take the data from January 2001 to November 2008 as out-of-sample (testing periods) data sets (95 observations), which is used to evaluate the good or bad performance of prediction based on some evaluation measurement. For evaluation, two typical indicators, normalized mean squared error (NMSE)[1] and directional statistics (D stat ) [1] are used. Given N pairs of the actual values (or targets, x t ) and predicted values (ˆx t ), the NMSE which normalizes the MSE by dividing it through the variance of respective series can be defined as NMSE = N t=1 (x t ˆx t ) 2 N t=1 (x t x t ) 2 = 1 1 N δ 2 N (x t ˆx t ) 2 (19) t=1 where δ 2 is the estimated variance of the data and x t the mean. Usually the NMSE is only a level prediction evaluation criterion, which is not enough for financial forecasting. For this, the directional statistics (D stat ) is developed, which is defined by D stat = 1 N a t 100% (20) N t=1 where a t =1 if (x t+1 x t )(ˆx t+1 x t ) 0, and a t =0 otherwise. In addition, for comparison purpose, a reverse model relative to C-ALSSVR, called the C-descending LSSVR (C-DLSSVR), standard LSSVR without using the most recent 60 data points (LSSVR-60), and standard LSSVR (LSSVR) are used here. If the prediction performance of the C-DLSSVR and LSSVR-60 is worse than that of LSSVR, the prior knowledge that the recent training data points can offer more information for modeling than the distant training data points will be confirmed clearer. In the experiment of C-ALSSVR, the Gaussian function is used as the kernel function and two ascending forms are examined. The regularization parameter C and kernel parameter σ are determined by the cross validation method. In particular, the validation set is used to choose the best combination of C, σ, and the optimal control rate r in the exponential ascending form. These values could
8 Foreign Exchange Rates Forecasting with a C-ALSSVR Model 613 vary in different foreign exchange rates due to different characteristics of foreign exchange rates. The LS-SVMlab1.5 for solving the regression problem [15] is implemented in this experiment and the program is developed using Matlab language. For the LSSVR, the LSSVR-60, and the C-descending LSSVR which will put more weights on the more distant training data points, the same settings with C-ALSSVR are adopted. Using the above experimental design, the corresponding computational results are reported in Tables 1 and 2 from the point of level prediction and direction prediction. In the two tables, a clear comparison of various methods for the three currencies is presented via NMSE and D stat. Generally speaking, the results obtained from the two tables also indicate that the prediction performance of the proposed C-ascending LSSVR model is better than those of the standard LSSVR, LSSVR-60 and C-DLSSVR models for the three main currencies. Table 1. The NMSE comparisons of different models for three foreign exchange rates Models GBP EUR JPY NMSE Rank NMSE Rank NMSE Rank LSSVR LSSVR C-DLSSVR Lin C-DLSSVR Exp C-ALSSVR Lin C-ALSSVR Exp Table 2. The D stat comparisons of different models for three foreign exchange rates Models GBP EUR JPY D stat(%) Rank D stat(%) Rank D stat(%) Rank LSSVR LSSVR C-DLSSVR Lin C-DLSSVR Exp C-ALSSVR Lin C-ALSSVR Exp Focusing on the NMSE indicator, our proposed C-ascending LSSVR model with exponential ascending form performs the best in all the cases, followed by the C-ascending LSSVR model with linear ascending form, standard LSSVR, LSSVR-60 model, and the C-descending LSSVR model is the worst. This indicates that the proposed C-ascending LSSVR model is more suitable for foreign exchange rates prediction than the standard LSSVR and C-descending LSSVR models. Interestingly, for the testing case of JPY, the NMSEs of the all models are larger than those of the other two currencies. This might be because the JPY is more volatile than the other two currencies.
9 614 L. Yu, X. Zhang, and S. Wang However, the low NMSE does not necessarily mean that there is a high hit ratio for foreign exchange movement direction prediction. Thus the D stat comparison is necessary for business practitioners. Focusing on D stat of Table 2, we are not hard to find that the proposed C-ascending LSSVR model outperforms the other three models according to the ranking; furthermore, from the business practitioners point of view, D stat is more important than NMSE because the former is an important decision criterion in foreign exchange trading. With reference to Table 2, the differences between the different models are very significant. For instance, for the EUR testing case, the D stat for the C-DLSSVR model with exponential form is only 55.79%, for the LSSVR-60 method it is 66.32%, and the D stat for the standard LSSVR model is 73.68%; while for the C-ALSSVR model with exponential form, D stat reaches 84.21%. Furthermore, like NMSE indicator, the proposed C-ALSSVR model with exponential form also performs the best in all the cases, followed by the C-ALSSVR model with linear form and standard LSSVR model, and the C-DLSSVR model with exponential form performs the worst. The main reasons are that in the foreign exchange rates forecasting the recent training data points are more important than the distant training data points. That is, the recent data could offer more information for modeling. By incorporating this prior knowledge into the LSSVR methods, the C-ALSSVR models are more effective in foreign exchange rates forecasting than the standard LSSVR models. 5 Concluding Remarks In this study, a new least squares support vector regression (LSSVR) model, called C-ascending LSSVR (C-ALSSVR) model, is proposed for foreign exchange rates prediction. In terms of the empirical results, we can find that across different models for the test cases of three main currencies British pounds (GBP), euros (EUR) and Japanese yen (JPY) on the basis of different evaluation criteria, our C-ascending LSSVR models perform the best. In the presented forecasting cases, the NMSE is the lowest and the D stat is the highest, indicating that the proposed C-ascending LSSVR models can be used as a promising tool for foreign exchange rates prediction. In addition, an interesting research direction is to explore more complex ascending functions which can closely follow the dynamics of foreign exchange rates series. Furthermore, the proposed C-ascending LSSVR models can be easily extended into other financial time series forecasting problems. We will look into these issues in the future research. Acknowledgements This work is partially supported by grants from the National Natural Science Foundation of China (NSFC No , ) and the Knowledge Innovation Program of the Chinese Academy of Sciences.
10 References Foreign Exchange Rates Forecasting with a C-ALSSVR Model Yu, L., Wang, S.Y., Lai, K.K.: Foreign-Exchange-Rate Forecasting with Artificial Neural Networks. Springer, New York (2007) 2. Lai, K.K., Yu, L., Wang, S.Y., Huang, W.: Hybridizing Exponential Smoothing and Neural Network for Financial Time Series Predication. In: Alexandrov, V.N., van Albada, G.D., Sloot, P.M.A., Dongarra, J. (eds.) ICCS LNCS, vol. 3994, pp Springer, Heidelberg (2006) 3. De Matos, G.: Neural Networks for Forecasting Exchange Rate. M. Sc. Thesis. The University of Manitoba, Canada (1994) 4. Kuan, C.M., Liu, T.: Forecasting Exchange Rates Using Feedforward and Recurrent Neural Networks. Journal of Applied Econometrics 10, (1995) 5. Tenti, P.: Forecasting Foreign Exchange Rates Using Recurrent Neural Networks. Applied Artificial Intelligence 10, (1996) 6. Hsu, W., Hsu, L.S., Tenorio, M.F.: A Neural Network Procedure for Selecting Predictive Indicators in Currency Trading. In: Refenes, A.N. (ed.) Neural Networks in the Capital Markets, pp John Wiley and Sons, New York (1995) 7. Leung, M.T., Chen, A.S., Daouk, H.: Forecasting Exchange Rates Using General Regression Neural Networks. Computers & Operations Research 27, (2000) 8. Chen, A.S., Leung, M.T.: Regression Neural Network for Error Correction in Foreign Exchange Rate Forecasting and Trading. Computers & Operations Research 31, (2004) 9. Yu, L., Wang, S.Y., Lai, K.K.: Adaptive Smoothing Neural Networks in Foreign Exchange Rate Forecasting. In: Sunderam, V.S., van Albada, G.D., Sloot, P.M.A., Dongarra, J. (eds.) ICCS LNCS, vol. 3516, pp Springer, Heidelberg (2005) 10. Vapnik, V.: The Nature of Statistical Learning Theory. Springer, New York (1995) 11. Tay, F.E.H., Cao, L.J.: Modified Support Vector Machines in Financial Time Series Forecasting. Neurocomputing 48, (2005) 12. Scholkopf, B., Burges, C., Smola, A. (eds.): Advances in Kernel Methods Support Vector Learning. MIT Press, Cambridge (1998) 13. Suykens, J.A.K., Lukas, L., Vandewalle, J.: Sparse Approximation Using Least Squares Support Vector Machines. In: Proceedings of The 2000 IEEE International Symposium on Circuits and Systems, vol. 2, pp IEEE Press, New York (2000) 14. Refenes, A.N., Bentz, Y., Bunn, D.W., Burgess, A.N., Zapranis, A.D.: Financial Time Series Modeling with Discounted Least Squares Back-Propagation. Neurocomputing 14, (1997) 15. Pelckmans, K., Suykens, J.A.K., Van Gestel, T., De Brabanter, J., Lukas, L., Hamers, B., De Moor, B., Vandewalle, J.: LS-SVMlab: a Matlab/C Toolbox for Least Squares Support Vector Machines. Internal Report 02-44, ESAT-SISTA, K.U. Leuven, Leuven, Belgium (2002)
Selection of the Appropriate Lag Structure of Foreign Exchange Rates Forecasting Based on Autocorrelation Coefficient
Selection of the Appropriate Lag Structure of Foreign Exchange Rates Forecasting Based on Autocorrelation Coefficient Wei Huang 1,2, Shouyang Wang 2, Hui Zhang 3,4, and Renbin Xiao 1 1 School of Management,
More informationDiscussion About Nonlinear Time Series Prediction Using Least Squares Support Vector Machine
Commun. Theor. Phys. (Beijing, China) 43 (2005) pp. 1056 1060 c International Academic Publishers Vol. 43, No. 6, June 15, 2005 Discussion About Nonlinear Time Series Prediction Using Least Squares Support
More informationSupport Vector Machine Regression for Volatile Stock Market Prediction
Support Vector Machine Regression for Volatile Stock Market Prediction Haiqin Yang, Laiwan Chan, and Irwin King Department of Computer Science and Engineering The Chinese University of Hong Kong Shatin,
More informationOutliers Treatment in Support Vector Regression for Financial Time Series Prediction
Outliers Treatment in Support Vector Regression for Financial Time Series Prediction Haiqin Yang, Kaizhu Huang, Laiwan Chan, Irwin King, and Michael R. Lyu Department of Computer Science and Engineering
More informationAn Improved Conjugate Gradient Scheme to the Solution of Least Squares SVM
An Improved Conjugate Gradient Scheme to the Solution of Least Squares SVM Wei Chu Chong Jin Ong chuwei@gatsby.ucl.ac.uk mpeongcj@nus.edu.sg S. Sathiya Keerthi mpessk@nus.edu.sg Control Division, Department
More informationNON-FIXED AND ASYMMETRICAL MARGIN APPROACH TO STOCK MARKET PREDICTION USING SUPPORT VECTOR REGRESSION. Haiqin Yang, Irwin King and Laiwan Chan
In The Proceedings of ICONIP 2002, Singapore, 2002. NON-FIXED AND ASYMMETRICAL MARGIN APPROACH TO STOCK MARKET PREDICTION USING SUPPORT VECTOR REGRESSION Haiqin Yang, Irwin King and Laiwan Chan Department
More informationSupport Vector Regression (SVR) Descriptions of SVR in this discussion follow that in Refs. (2, 6, 7, 8, 9). The literature
Support Vector Regression (SVR) Descriptions of SVR in this discussion follow that in Refs. (2, 6, 7, 8, 9). The literature suggests the design variables should be normalized to a range of [-1,1] or [0,1].
More informationForecast daily indices of solar activity, F10.7, using support vector regression method
Research in Astron. Astrophys. 9 Vol. 9 No. 6, 694 702 http://www.raa-journal.org http://www.iop.org/journals/raa Research in Astronomy and Astrophysics Forecast daily indices of solar activity, F10.7,
More informationModel Selection for LS-SVM : Application to Handwriting Recognition
Model Selection for LS-SVM : Application to Handwriting Recognition Mathias M. Adankon and Mohamed Cheriet Synchromedia Laboratory for Multimedia Communication in Telepresence, École de Technologie Supérieure,
More informationA GENERAL FORMULATION FOR SUPPORT VECTOR MACHINES. Wei Chu, S. Sathiya Keerthi, Chong Jin Ong
A GENERAL FORMULATION FOR SUPPORT VECTOR MACHINES Wei Chu, S. Sathiya Keerthi, Chong Jin Ong Control Division, Department of Mechanical Engineering, National University of Singapore 0 Kent Ridge Crescent,
More informationEconomic Modelling 29 (2012) Contents lists available at SciVerse ScienceDirect. Economic Modelling
Economic Modelling 29 (2012) 2583 2590 Contents lists available at SciVerse ScienceDirect Economic Modelling journal homepage: www.elsevier.com/locate/ecmod Empirical mode decomposition based least squares
More informationPredict GARCH Based Volatility of Shanghai Composite Index by Recurrent Relevant Vector Machines and Recurrent Least Square Support Vector Machines
Predict GARCH Based Volatility of Shanghai Composite Index by Recurrent Relevant Vector Machines and Recurrent Least Square Support Vector Machines Phichhang Ou (Corresponding author) School of Business,
More informationFast Bootstrap for Least-square Support Vector Machines
ESA'2004 proceedings - European Symposium on Artificial eural etworks Bruges (Belgium), 28-30 April 2004, d-side publi., ISB 2-930307-04-8, pp. 525-530 Fast Bootstrap for Least-square Support Vector Machines
More informationResearch Article Time Series Adaptive Online Prediction Method Combined with Modified LS-SVR and AGO
Mathematical Problems in Engineering Volume 2012, Article ID 985930, 12 pages doi:10.1155/2012/985930 Research Article Time Series Adaptive Online Prediction Method Combined with Modified LS-SVR and AGO
More informationLearning with kernels and SVM
Learning with kernels and SVM Šámalova chata, 23. května, 2006 Petra Kudová Outline Introduction Binary classification Learning with Kernels Support Vector Machines Demo Conclusion Learning from data find
More informationDiscussion of Some Problems About Nonlinear Time Series Prediction Using ν-support Vector Machine
Commun. Theor. Phys. (Beijing, China) 48 (2007) pp. 117 124 c International Academic Publishers Vol. 48, No. 1, July 15, 2007 Discussion of Some Problems About Nonlinear Time Series Prediction Using ν-support
More informationLinear & nonlinear classifiers
Linear & nonlinear classifiers Machine Learning Hamid Beigy Sharif University of Technology Fall 1396 Hamid Beigy (Sharif University of Technology) Linear & nonlinear classifiers Fall 1396 1 / 44 Table
More informationChapter 9. Support Vector Machine. Yongdai Kim Seoul National University
Chapter 9. Support Vector Machine Yongdai Kim Seoul National University 1. Introduction Support Vector Machine (SVM) is a classification method developed by Vapnik (1996). It is thought that SVM improved
More informationKalman Filter and SVR Combinations in Forecasting US Unemployment
Kalman Filter and SVR Combinations in Forecasting US Unemployment Georgios Sermpinis 1, Charalampos Stasinakis 1, and Andreas Karathanasopoulos 2 1 University of Glasgow Business School georgios.sermpinis@glasgow.ac.uk,
More informationClassifier Complexity and Support Vector Classifiers
Classifier Complexity and Support Vector Classifiers Feature 2 6 4 2 0 2 4 6 8 RBF kernel 10 10 8 6 4 2 0 2 4 6 Feature 1 David M.J. Tax Pattern Recognition Laboratory Delft University of Technology D.M.J.Tax@tudelft.nl
More informationRAINFALL RUNOFF MODELING USING SUPPORT VECTOR REGRESSION AND ARTIFICIAL NEURAL NETWORKS
CEST2011 Rhodes, Greece Ref no: XXX RAINFALL RUNOFF MODELING USING SUPPORT VECTOR REGRESSION AND ARTIFICIAL NEURAL NETWORKS D. BOTSIS1 1, P. LATINOPOULOS 2 and K. DIAMANTARAS 3 1&2 Department of Civil
More informationSVM TRADE-OFF BETWEEN MAXIMIZE THE MARGIN AND MINIMIZE THE VARIABLES USED FOR REGRESSION
International Journal of Pure and Applied Mathematics Volume 87 No. 6 2013, 741-750 ISSN: 1311-8080 (printed version); ISSN: 1314-3395 (on-line version) url: http://www.ijpam.eu doi: http://dx.doi.org/10.12732/ijpam.v87i6.2
More informationSupport Vector Machine (continued)
Support Vector Machine continued) Overlapping class distribution: In practice the class-conditional distributions may overlap, so that the training data points are no longer linearly separable. We need
More informationScale-Invariance of Support Vector Machines based on the Triangular Kernel. Abstract
Scale-Invariance of Support Vector Machines based on the Triangular Kernel François Fleuret Hichem Sahbi IMEDIA Research Group INRIA Domaine de Voluceau 78150 Le Chesnay, France Abstract This paper focuses
More informationLinear & nonlinear classifiers
Linear & nonlinear classifiers Machine Learning Hamid Beigy Sharif University of Technology Fall 1394 Hamid Beigy (Sharif University of Technology) Linear & nonlinear classifiers Fall 1394 1 / 34 Table
More informationDo we need Experts for Time Series Forecasting?
Do we need Experts for Time Series Forecasting? Christiane Lemke and Bogdan Gabrys Bournemouth University - School of Design, Engineering and Computing Poole House, Talbot Campus, Poole, BH12 5BB - United
More informationSupport Vector Machines
Support Vector Machines Stephan Dreiseitl University of Applied Sciences Upper Austria at Hagenberg Harvard-MIT Division of Health Sciences and Technology HST.951J: Medical Decision Support Overview Motivation
More informationKNOWLEDGE-BASED SUPPORT VECTOR SYNTHESIS OF THE MICROSTRIP LINES
Progress In Electromagnetics Research, PIER 92, 65 77, 2009 KNOWLEDGE-BASED SUPPORT VECTOR SYNTHESIS OF THE MICROSTRIP LINES N. T. Tokan and F. Güneş Department of Electronics and Communication Engineering
More informationPredicting Time of Peak Foreign Exchange Rates. Charles Mulemi, Lucio Dery 0. ABSTRACT
Predicting Time of Peak Foreign Exchange Rates Charles Mulemi, Lucio Dery 0. ABSTRACT This paper explores various machine learning models of predicting the day foreign exchange rates peak in a given window.
More informationECE662: Pattern Recognition and Decision Making Processes: HW TWO
ECE662: Pattern Recognition and Decision Making Processes: HW TWO Purdue University Department of Electrical and Computer Engineering West Lafayette, INDIANA, USA Abstract. In this report experiments are
More informationContent. Learning. Regression vs Classification. Regression a.k.a. function approximation and Classification a.k.a. pattern recognition
Content Andrew Kusiak Intelligent Systems Laboratory 239 Seamans Center The University of Iowa Iowa City, IA 52242-527 andrew-kusiak@uiowa.edu http://www.icaen.uiowa.edu/~ankusiak Introduction to learning
More informationCompactly supported RBF kernels for sparsifying the Gram matrix in LS-SVM regression models
Compactly supported RBF kernels for sparsifying the Gram matrix in LS-SVM regression models B. Hamers, J.A.K. Suykens, B. De Moor K.U.Leuven, ESAT-SCD/SISTA, Kasteelpark Arenberg, B-3 Leuven, Belgium {bart.hamers,johan.suykens}@esat.kuleuven.ac.be
More informationLearning Kernel Parameters by using Class Separability Measure
Learning Kernel Parameters by using Class Separability Measure Lei Wang, Kap Luk Chan School of Electrical and Electronic Engineering Nanyang Technological University Singapore, 3979 E-mail: P 3733@ntu.edu.sg,eklchan@ntu.edu.sg
More informationA Support Vector Regression Model for Forecasting Rainfall
A Support Vector Regression for Forecasting Nasimul Hasan 1, Nayan Chandra Nath 1, Risul Islam Rasel 2 Department of Computer Science and Engineering, International Islamic University Chittagong, Bangladesh
More informationElectrical Energy Modeling In Y2E2 Building Based On Distributed Sensors Information
Electrical Energy Modeling In Y2E2 Building Based On Distributed Sensors Information Mahmoud Saadat Saman Ghili Introduction Close to 40% of the primary energy consumption in the U.S. comes from commercial
More informationNew Least Squares Support Vector Machines Based on Matrix Patterns
Neural Processing Letters (2007) 26:41 56 DOI 10.1007/s11063-007-9041-1 New Least Squares Support Vector Machines Based on Matrix Patterns Zhe Wang Songcan Chen Received: 14 August 2005 / Accepted: 7 May
More informationSUPPORT VECTOR REGRESSION WITH A GENERALIZED QUADRATIC LOSS
SUPPORT VECTOR REGRESSION WITH A GENERALIZED QUADRATIC LOSS Filippo Portera and Alessandro Sperduti Dipartimento di Matematica Pura ed Applicata Universit a di Padova, Padova, Italy {portera,sperduti}@math.unipd.it
More informationAn introduction to Support Vector Machines
1 An introduction to Support Vector Machines Giorgio Valentini DSI - Dipartimento di Scienze dell Informazione Università degli Studi di Milano e-mail: valenti@dsi.unimi.it 2 Outline Linear classifiers
More informationLinear Dependency Between and the Input Noise in -Support Vector Regression
544 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 14, NO. 3, MAY 2003 Linear Dependency Between the Input Noise in -Support Vector Regression James T. Kwok Ivor W. Tsang Abstract In using the -support vector
More informationSARIMA-ELM Hybrid Model for Forecasting Tourist in Nepal
Volume-03 Issue-07 July-2018 ISSN: 2455-3085 (Online) www.rrjournals.com [UGC Listed Journal] SARIMA-ELM Hybrid Model for Forecasting Tourist in Nepal *1 Kadek Jemmy Waciko & 2 Ismail B *1 Research Scholar,
More informationDiscriminative Models
No.5 Discriminative Models Hui Jiang Department of Electrical Engineering and Computer Science Lassonde School of Engineering York University, Toronto, Canada Outline Generative vs. Discriminative models
More informationIteratively Reweighted Least Square for Asymmetric L 2 -Loss Support Vector Regression
Iteratively Reweighted Least Square for Asymmetric L 2 -Loss Support Vector Regression Songfeng Zheng Department of Mathematics Missouri State University Springfield, MO 65897 SongfengZheng@MissouriState.edu
More informationBits of Machine Learning Part 1: Supervised Learning
Bits of Machine Learning Part 1: Supervised Learning Alexandre Proutiere and Vahan Petrosyan KTH (The Royal Institute of Technology) Outline of the Course 1. Supervised Learning Regression and Classification
More informationA Tutorial on Support Vector Machine
A Tutorial on School of Computing National University of Singapore Contents Theory on Using with Other s Contents Transforming Theory on Using with Other s What is a classifier? A function that maps instances
More informationDiscriminative Direction for Kernel Classifiers
Discriminative Direction for Kernel Classifiers Polina Golland Artificial Intelligence Lab Massachusetts Institute of Technology Cambridge, MA 02139 polina@ai.mit.edu Abstract In many scientific and engineering
More informationSUPPORT VECTOR MACHINE FOR THE SIMULTANEOUS APPROXIMATION OF A FUNCTION AND ITS DERIVATIVE
SUPPORT VECTOR MACHINE FOR THE SIMULTANEOUS APPROXIMATION OF A FUNCTION AND ITS DERIVATIVE M. Lázaro 1, I. Santamaría 2, F. Pérez-Cruz 1, A. Artés-Rodríguez 1 1 Departamento de Teoría de la Señal y Comunicaciones
More informationSolving optimal control problems by PSO-SVM
Computational Methods for Differential Equations http://cmde.tabrizu.ac.ir Vol. 6, No. 3, 2018, pp. 312-325 Solving optimal control problems by PSO-SVM Elham Salehpour Department of Mathematics, Nowshahr
More informationFORECASTING CHINA S FOREIGN TRADE VOLUME WITH A KERNEL-BASED HYBRID ECONOMETRIC-AI ENSEMBLE LEARNING APPROACH
Jrl Syst Sci & Complexity (2008) 21: 1 19 FORECASTING CHINA S FOREIGN TRADE VOLUME WITH A KERNEL-BASED HYBRID ECONOMETRIC-AI ENSEMBLE LEARNING APPROACH Lean YU Shouyang WANG Kin Keung LAI Received: 20
More informationTransformer Top-Oil Temperature Modeling and Simulation
Transformer Top-Oil Temperature Modeling and Simulation T. C. B. N. Assunção, J. L. Silvino, and P. Resende Abstract The winding hot-spot temperature is one of the most critical parameters that affect
More informationOn-line Support Vector Machines and Optimization Strategies
On-line Support Vector Machines and Optimization Strategies J.M. Górriz, C.G. Puntonet, M. Salmerón, R. Martin-Clemente and S. Hornillo-Mellado and J. Ortega E.S.I., Informática, Universidad de Granada
More informationSupport Vector Machine for Classification and Regression
Support Vector Machine for Classification and Regression Ahlame Douzal AMA-LIG, Université Joseph Fourier Master 2R - MOSIG (2013) November 25, 2013 Loss function, Separating Hyperplanes, Canonical Hyperplan
More informationRobust Kernel-Based Regression
Robust Kernel-Based Regression Budi Santosa Department of Industrial Engineering Sepuluh Nopember Institute of Technology Kampus ITS Surabaya Surabaya 60111,Indonesia Theodore B. Trafalis School of Industrial
More informationThe Neural Support Vector Machine
The Neural Support Vector Machine M.A. Wiering a M.H. van der Ree a M.J. Embrechts b M.F. Stollenga c A. Meijster a A. Nolte d L.R.B. Schomaker a a Institute of Artificial Intelligence and Cognitive Engineering,
More informationReading, UK 1 2 Abstract
, pp.45-54 http://dx.doi.org/10.14257/ijseia.2013.7.5.05 A Case Study on the Application of Computational Intelligence to Identifying Relationships between Land use Characteristics and Damages caused by
More informationGaussian Fitting Based FDA for Chemometrics
Gaussian Fitting Based FDA for Chemometrics Tuomas Kärnä and Amaury Lendasse Helsinki University of Technology, Laboratory of Computer and Information Science, P.O. Box 54 FI-25, Finland tuomas.karna@hut.fi,
More informationSupport Vector Regression with Automatic Accuracy Control B. Scholkopf y, P. Bartlett, A. Smola y,r.williamson FEIT/RSISE, Australian National University, Canberra, Australia y GMD FIRST, Rudower Chaussee
More informationDiscriminative Models
No.5 Discriminative Models Hui Jiang Department of Electrical Engineering and Computer Science Lassonde School of Engineering York University, Toronto, Canada Outline Generative vs. Discriminative models
More informationSupport Vector Machine & Its Applications
Support Vector Machine & Its Applications A portion (1/3) of the slides are taken from Prof. Andrew Moore s SVM tutorial at http://www.cs.cmu.edu/~awm/tutorials Mingyue Tan The University of British Columbia
More informationIterative ARIMA-Multiple Support Vector Regression models for long term time series prediction
and Machine Learning Bruges (Belgium), 23-25 April 24, i6doccom publ, ISBN 978-2874995-7 Available from http://wwwi6doccom/fr/livre/?gcoi=2843244 Iterative ARIMA-Multiple Support Vector Regression models
More informationConstrained Optimization and Support Vector Machines
Constrained Optimization and Support Vector Machines Man-Wai MAK Dept. of Electronic and Information Engineering, The Hong Kong Polytechnic University enmwmak@polyu.edu.hk http://www.eie.polyu.edu.hk/
More informationLinear vs Non-linear classifier. CS789: Machine Learning and Neural Network. Introduction
Linear vs Non-linear classifier CS789: Machine Learning and Neural Network Support Vector Machine Jakramate Bootkrajang Department of Computer Science Chiang Mai University Linear classifier is in the
More informationData Mining. Linear & nonlinear classifiers. Hamid Beigy. Sharif University of Technology. Fall 1396
Data Mining Linear & nonlinear classifiers Hamid Beigy Sharif University of Technology Fall 1396 Hamid Beigy (Sharif University of Technology) Data Mining Fall 1396 1 / 31 Table of contents 1 Introduction
More informationSupport Vector Machines. CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington
Support Vector Machines CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington 1 A Linearly Separable Problem Consider the binary classification
More informationLeast Squares SVM Regression
Least Squares SVM Regression Consider changing SVM to LS SVM by making following modifications: min (w,e) ½ w 2 + ½C Σ e(i) 2 subject to d(i) (w T Φ( x(i))+ b) = e(i), i, and C>0. Note that e(i) is error
More informationNeural Networks. Prof. Dr. Rudolf Kruse. Computational Intelligence Group Faculty for Computer Science
Neural Networks Prof. Dr. Rudolf Kruse Computational Intelligence Group Faculty for Computer Science kruse@iws.cs.uni-magdeburg.de Rudolf Kruse Neural Networks 1 Supervised Learning / Support Vector Machines
More informationIntroduction to Support Vector Machines
Introduction to Support Vector Machines Hsuan-Tien Lin Learning Systems Group, California Institute of Technology Talk in NTU EE/CS Speech Lab, November 16, 2005 H.-T. Lin (Learning Systems Group) Introduction
More informationSupport Vector Machines
Support Vector Machines Le Song Machine Learning I CSE 6740, Fall 2013 Naïve Bayes classifier Still use Bayes decision rule for classification P y x = P x y P y P x But assume p x y = 1 is fully factorized
More informationLeast Absolute Shrinkage is Equivalent to Quadratic Penalization
Least Absolute Shrinkage is Equivalent to Quadratic Penalization Yves Grandvalet Heudiasyc, UMR CNRS 6599, Université de Technologie de Compiègne, BP 20.529, 60205 Compiègne Cedex, France Yves.Grandvalet@hds.utc.fr
More informationLocal Support Vector Regression For Financial Time Series Prediction
26 International Joint Conference on eural etworks Sheraton Vancouver Wall Centre Hotel, Vancouver, BC, Canada July 6-2, 26 Local Support Vector Regression For Financial Time Series Prediction Kaizhu Huang,
More informationDeep Learning Architecture for Univariate Time Series Forecasting
CS229,Technical Report, 2014 Deep Learning Architecture for Univariate Time Series Forecasting Dmitry Vengertsev 1 Abstract This paper studies the problem of applying machine learning with deep architecture
More informationNeural Networks and the Back-propagation Algorithm
Neural Networks and the Back-propagation Algorithm Francisco S. Melo In these notes, we provide a brief overview of the main concepts concerning neural networks and the back-propagation algorithm. We closely
More informationDirect and Recursive Prediction of Time Series Using Mutual Information Selection
Direct and Recursive Prediction of Time Series Using Mutual Information Selection Yongnan Ji, Jin Hao, Nima Reyhani, and Amaury Lendasse Neural Network Research Centre, Helsinki University of Technology,
More informationFinding Robust Solutions to Dynamic Optimization Problems
Finding Robust Solutions to Dynamic Optimization Problems Haobo Fu 1, Bernhard Sendhoff, Ke Tang 3, and Xin Yao 1 1 CERCIA, School of Computer Science, University of Birmingham, UK Honda Research Institute
More informationSupport Vector Ordinal Regression using Privileged Information
Support Vector Ordinal Regression using Privileged Information Fengzhen Tang 1, Peter Tiňo 2, Pedro Antonio Gutiérrez 3 and Huanhuan Chen 4 1,2,4- The University of Birmingham, School of Computer Science,
More informationVC dimension, Model Selection and Performance Assessment for SVM and Other Machine Learning Algorithms
03/Feb/2010 VC dimension, Model Selection and Performance Assessment for SVM and Other Machine Learning Algorithms Presented by Andriy Temko Department of Electrical and Electronic Engineering Page 2 of
More informationOutline. Basic concepts: SVM and kernels SVM primal/dual problems. Chih-Jen Lin (National Taiwan Univ.) 1 / 22
Outline Basic concepts: SVM and kernels SVM primal/dual problems Chih-Jen Lin (National Taiwan Univ.) 1 / 22 Outline Basic concepts: SVM and kernels Basic concepts: SVM and kernels SVM primal/dual problems
More informationMachine Learning. Support Vector Machines. Manfred Huber
Machine Learning Support Vector Machines Manfred Huber 2015 1 Support Vector Machines Both logistic regression and linear discriminant analysis learn a linear discriminant function to separate the data
More informationPattern Matching and Neural Networks based Hybrid Forecasting System
Pattern Matching and Neural Networks based Hybrid Forecasting System Sameer Singh and Jonathan Fieldsend PA Research, Department of Computer Science, University of Exeter, Exeter, UK Abstract In this paper
More informationFormulation with slack variables
Formulation with slack variables Optimal margin classifier with slack variables and kernel functions described by Support Vector Machine (SVM). min (w,ξ) ½ w 2 + γσξ(i) subject to ξ(i) 0 i, d(i) (w T x(i)
More informationSpace-Time Kernels. Dr. Jiaqiu Wang, Dr. Tao Cheng James Haworth University College London
Space-Time Kernels Dr. Jiaqiu Wang, Dr. Tao Cheng James Haworth University College London Joint International Conference on Theory, Data Handling and Modelling in GeoSpatial Information Science, Hong Kong,
More informationMultiple Similarities Based Kernel Subspace Learning for Image Classification
Multiple Similarities Based Kernel Subspace Learning for Image Classification Wang Yan, Qingshan Liu, Hanqing Lu, and Songde Ma National Laboratory of Pattern Recognition, Institute of Automation, Chinese
More informationEnsembles of Nearest Neighbor Forecasts
Ensembles of Nearest Neighbor Forecasts Dragomir Yankov 1, Dennis DeCoste 2, and Eamonn Keogh 1 1 University of California, Riverside CA 92507, USA, {dyankov,eamonn}@cs.ucr.edu, 2 Yahoo! Research, 3333
More informationCheng Soon Ong & Christian Walder. Canberra February June 2018
Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 Outlines Overview Introduction Linear Algebra Probability Linear Regression
More informationStable Adaptive Momentum for Rapid Online Learning in Nonlinear Systems
Stable Adaptive Momentum for Rapid Online Learning in Nonlinear Systems Thore Graepel and Nicol N. Schraudolph Institute of Computational Science ETH Zürich, Switzerland {graepel,schraudo}@inf.ethz.ch
More informationReview: Support vector machines. Machine learning techniques and image analysis
Review: Support vector machines Review: Support vector machines Margin optimization min (w,w 0 ) 1 2 w 2 subject to y i (w 0 + w T x i ) 1 0, i = 1,..., n. Review: Support vector machines Margin optimization
More informationSupport'Vector'Machines. Machine(Learning(Spring(2018 March(5(2018 Kasthuri Kannan
Support'Vector'Machines Machine(Learning(Spring(2018 March(5(2018 Kasthuri Kannan kasthuri.kannan@nyumc.org Overview Support Vector Machines for Classification Linear Discrimination Nonlinear Discrimination
More informationPredict Time Series with Multiple Artificial Neural Networks
, pp. 313-324 http://dx.doi.org/10.14257/ijhit.2016.9.7.28 Predict Time Series with Multiple Artificial Neural Networks Fei Li 1, Jin Liu 1 and Lei Kong 2,* 1 College of Information Engineering, Shanghai
More informationSupport Vector Machines
Support Vector Machines Tobias Pohlen Selected Topics in Human Language Technology and Pattern Recognition February 10, 2014 Human Language Technology and Pattern Recognition Lehrstuhl für Informatik 6
More informationML (cont.): SUPPORT VECTOR MACHINES
ML (cont.): SUPPORT VECTOR MACHINES CS540 Bryan R Gibson University of Wisconsin-Madison Slides adapted from those used by Prof. Jerry Zhu, CS540-1 1 / 40 Support Vector Machines (SVMs) The No-Math Version
More informationMachine Learning. Lecture 9: Learning Theory. Feng Li.
Machine Learning Lecture 9: Learning Theory Feng Li fli@sdu.edu.cn https://funglee.github.io School of Computer Science and Technology Shandong University Fall 2018 Why Learning Theory How can we tell
More informationCIS 520: Machine Learning Oct 09, Kernel Methods
CIS 520: Machine Learning Oct 09, 207 Kernel Methods Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture They may or may not cover all the material discussed
More informationSupport Vector Machines (SVM) in bioinformatics. Day 1: Introduction to SVM
1 Support Vector Machines (SVM) in bioinformatics Day 1: Introduction to SVM Jean-Philippe Vert Bioinformatics Center, Kyoto University, Japan Jean-Philippe.Vert@mines.org Human Genome Center, University
More informationHybrid particle swarm algorithm for solving nonlinear constraint. optimization problem [5].
Hybrid particle swarm algorithm for solving nonlinear constraint optimization problems BINGQIN QIAO, XIAOMING CHANG Computers and Software College Taiyuan University of Technology Department of Economic
More informationNonlinear System Identification using Support Vector Regression
Nonlinear System Identification using Support Vector Regression Saneej B.C. PhD Student Department of Chemical and Materials Engineering University of Alberta Outline 2 1. Objectives 2. Nonlinearity in
More informationMulti-Model Integration for Long-Term Time Series Prediction
Multi-Model Integration for Long-Term Time Series Prediction Zifang Huang, Mei-Ling Shyu, James M. Tien Department of Electrical and Computer Engineering University of Miami, Coral Gables, FL, USA z.huang3@umiami.edu,
More informationSupport Vector Machine (SVM) & Kernel CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2012
Support Vector Machine (SVM) & Kernel CE-717: Machine Learning Sharif University of Technology M. Soleymani Fall 2012 Linear classifier Which classifier? x 2 x 1 2 Linear classifier Margin concept x 2
More informationCS798: Selected topics in Machine Learning
CS798: Selected topics in Machine Learning Support Vector Machine Jakramate Bootkrajang Department of Computer Science Chiang Mai University Jakramate Bootkrajang CS798: Selected topics in Machine Learning
More informationStatistical Machine Learning from Data
Samy Bengio Statistical Machine Learning from Data 1 Statistical Machine Learning from Data Support Vector Machines Samy Bengio IDIAP Research Institute, Martigny, Switzerland, and Ecole Polytechnique
More informationLMS Algorithm Summary
LMS Algorithm Summary Step size tradeoff Other Iterative Algorithms LMS algorithm with variable step size: w(k+1) = w(k) + µ(k)e(k)x(k) When step size µ(k) = µ/k algorithm converges almost surely to optimal
More information