Duality, Geometry, and Support Vector Regression
|
|
- Edmund Marshall
- 5 years ago
- Views:
Transcription
1 ualit, Geometr, and Support Vector Regression Jinbo Bi and Kristin P. Bennett epartment of Mathematical Sciences Rensselaer Poltechnic Institute Tro, NY 80 Abstract We develop an intuitive geometric framework for support vector regression (SVR). B eamining when ɛ-tubes eist, we show that SVR can be regarded as a classification problem in the dual space. Hard and soft ɛ-tubes are constructed b separating the conve or reduced conve hulls respectivel of the training data with the response variable shifted up and down b ɛ. A novel SVR model is proposed based on choosing the ma-margin plane between the two shifted datasets. Maimizing the margin corresponds to shrinking the effective ɛ-tube. In the proposed approach the effects of the choices of all parameters become clear geometricall. Introduction Support Vector Machines (SVMs) [6] are a ver robust methodolog for inference with minimal parameter choices. Intuitive geometric formulations eist for the classification case addressing both the error metric and capacit control [, ]. For linearl separable classification, the primal SVM finds the separating plane with maimum hard margin between two sets. The equivalent dual SVM computes the closest points in the conve hulls of the data from each class. For the inseparable case, the primal SVM optimizes the soft margin of separation between the two classes. The corresponding dual SVM finds the closest points in the reduced conve hulls. In this paper, we derive analogous arguments for SVM regression (SVR). We provide a geometric eplanation for SVR with the ɛ-insensitive loss function. From the primal perspective, a linear function with no residuals greater than ɛ corresponds to an ɛ-tube constructed about the data in the space of the data attributes and the response variable [6] (see e.g. Figure (a)). The primar contribution of this work is a novel geometric interpretation of SVR from the dual perspective along with a mathematicall rigorous derivation of the geometric concepts. In Section, for a fied ɛ > 0 we eamine the question When does a perfect or hard
2 ɛ-tube eist?. With dualit analsis, the eistence of a hard ɛ-tube depends on the separabilit of two sets. The two sets consist of the training data augmented with the response variable shifted up and down b ɛ. In the dual space, regression becomes the classification problem of distinguishing between these two sets. The geometric formulations developed for the classification case [] become applicable to the regression case. We call the resulting formulation conve SVR (C-SVR) since it is based on conve hulls of the augmented training data. Much like in SVM classification, to compute a hard ɛ-tube, C-SVR computes the nearest points in the conve hulls of the augmented classes. The corresponding maimum margin (ma-margin) planes define the effective ɛ-tube. The size of margin determines how much the effective ɛ-tube shrinks. Similarl, to compute a soft ɛ-tube, reduced-conve SVR (RC-SVR) finds the closest points in the reduced conve hulls of the two augmented sets. This paper introduces the geometricall intuitive RC-SVR formulation which is a variation of the classic ɛ-svr [6] and ν-svr models [5]. If parameters are properl tuned, the methods perform similarl although not necessaril identicall. RC- SVR eliminates the pesk parameter C used in ɛ-svr and ν-svr. The geometric role or interpretation of C is not known for these formulations. The geometric roles of the two parameters of RC-SVR, ν and ɛ, are ver clear, facilitating model selection, especiall for noneperts. Like ν-svr, RC-SVR shrinks the ɛ-tube and has a parameter ν controlling the robustness of the solution. The parameter ɛ acts as an upper bound on the size of the allowable ɛ-insensitive error function. In addition, RC-SVR can be solved b fast and scalable nearest-point algorithms such as those used in [3] for SVM classification. When does a hard ɛ-tube eist? ε ε (a) - (b) - (c) - (d) Figure : The (a) primal hard ɛ 0-tube, and dual cases: (b) dual strictl separable ɛ > ɛ 0, (c) dual separable ɛ = ɛ 0, and (d) dual inseparable ɛ < ɛ 0. SVR constructs a regression model that minimizes some empirical risk measure regularized to control capacit. Let be the n predictor variables and the dependent response variable. In [6], Vapnik proposed using the ɛ-insensitive loss function L ɛ (,, f) = f() ɛ = ma (0, f() ɛ), in which an eample is in error if its residual f() is greater than ɛ. Plotting the points in (, ) space as in Figure (a), we see that for a perfect regression model the data fall in a hard ɛ-tube about the regression line. Let (X i, i ) be an eample where i =,,, m, X i is the i th predictor vector, and i is its response. The training data are then (X, ) where X i is a row of the matri X R m n and R m is the response. A hard ɛ-tube for a fied ɛ > 0 is defined as a plane = w + b satisfing ɛe Xw be ɛe where e is an m-dimensional vector of ones. When does a hard ɛ-tube eist? Clearl, for ɛ large enough such a tube alwas
3 eists for finite data. The smallest tube, the ɛ 0 -tube, can be found b optimizing: min ɛ s.t. ɛe Xw be ɛe () w,b,ɛ Note that the smallest tube is tpicall not the ɛ-svr solution. Let + and be formed b augmenting the data with the response variable respectivel increased and decreased b ɛ, i.e. + = {(X i, i + ɛ), i =,, m} and = {(X i, i ɛ), i =,, m}. Consider the simple problem in Figure (a). For an fied ɛ > 0, there are three possible cases: ɛ > ɛ 0 in which strict hard ɛ-tubes eist, ɛ = ɛ 0 in which onl ɛ 0 -tubes eist, and ɛ < ɛ 0 in which no hard ɛ-tubes eist. A strict hard ɛ-tube with no points on the edges of the tube onl eists for ɛ > ɛ 0. Figure (b-d) illustrates what happens in the dual space for each case. The conve hulls of + and are drawn along with the ma-margin plane in (b) and the supporting plane in (c) for separating the conve hulls. Clearl, the eistence of the tube is directl related to the separabilit of + and. If ɛ > ɛ 0 then a strict tube eists and the conve hulls of + and are strictl separable. There are infinitel man possible ɛ-tubes when ɛ > ɛ 0. One can see that the ma-margin plane separating + and corresponds to one such ɛ. In fact this plane forms an ˆɛ tube where ɛ > ˆɛ ɛ 0. If ɛ = ɛ 0, then the conve hulls of + and are separable but not strictl separable. The plane that separates the two conve hulls forms the ɛ 0 tube. In the last case, where ɛ < ɛ 0, the two sets + and intersect. No ɛ-tubes or ma-margin planes eist. It is eas to show b construction that if a hard ɛ-tube eists for a given ɛ > 0 then the conve hulls of + and will be separable. If a hard ɛ-tube eists, then there eists (w, b) such that ( + ɛe) Xw be 0, ( ɛe) Xw be 0. () For an conve combination of +, ( ) X (+ɛe) u where e u =, u 0 of points (X i, i + ɛ), i =,,, m, we have ( + ɛe) u w (X u) b 0. Similarl for, ( ) X ( ɛe) v where e v =, v 0 of points (X i, i ɛ), i =,,, m, we have ( ɛe) v w (X v) b 0. Then the plane = w + b in the ɛ-tube separates the two conve hulls. Note the separating plane and the ɛ-tube plane are the same. If no separating plane eists, then there is no tube. Gale s Theorem of the alternative can be used to precisel characterize the ɛ-tube. Theorem. (Conditions for eistence of hard ɛ-tube) A hard ɛ-tube eists for a given ɛ > 0 if and onl if the following sstem in (u, v) has no solution: X u = X v, e u = e v =, ( + ɛe) u ( ɛe) v < 0, u 0, v 0. (3) Proof A hard ɛ-tube eists if and onl if Sstem () has a solution. B Gale s Theorem of the alternative [4], sstem () has a solution if and onl if the following alternative sstem has no solution: X u = X v, e u = e v, ( + ɛe) u ( ɛe) v =, u 0, v 0. Rescaling b σ where σ = e u = e v > 0 ields the result. We use the following definitions of separation of conve sets. Let + and be nonempt conve sets. A plane H = { : w = α} is said to separate + and if w α, + and w α,. H is said to strictl separate + and if w α + for +, and w α for each where is a positive scalar. The sstem A c has a (or has no) solution if and onl if the alternative sstem A = 0, c =, 0 has no (or has a) solution.
4 Note that if ɛ ɛ 0 then ( + ɛe) u ( ɛe) v 0. for an (u, v) such that X u = X v, e u = e v =, u, v 0. So as a consequence of this theorem, if + and are separable, then a hard ɛ-tube eists. 3 Constructing the ɛ-tube For an ɛ > ɛ 0 infinitel man possible ɛ-tubes eist. Which ɛ-tube should be used? The linear program () can be solved to find the smallest ɛ 0 -tube. But this corresponds to just doing empirical risk minimization and ma result in poor generalization due to overfitting. We know capacit control or structural risk minimization is fundamental to the success of SVM classification and regression. We take our inspiration from SVM classification. In hard-margin SVM classification, the dual SVM formulation constructs the ma-margin plane b finding the two nearest points in the conve hulls of the two classes. The ma-margin plane is the plane bisecting these two points. We know that the eistence of the tube is linked to the separabilit of the shifted sets, + and. The ke insight is that the regression problem can be regarded as a classification problem between + and. The two sets + and defined as in Section both contain the same number of data points. The onl significant difference occurs along the dimension as the response variable is shifted up b ɛ in + and down b ɛ in. For ɛ > ɛ 0, the ma-margin separating plane corresponds to a hard ˆɛ-tube where ɛ > ˆɛ ɛ 0. The resulting tube is smaller than ɛ but not necessaril the smallest tube. Figure (b) shows the ma-margin plane found for ɛ > ɛ 0. Figure (a) shows that the corresponding linear regression function for this simple eample turns out to be the ɛ 0 tube. As in classification, we will have a hard and soft ɛ-tube case. The soft ɛ-tube with ɛ ɛ 0 is used to obtain good generalization when there are outliers. 3. The hard ɛ-tube case We now appl the dual conve hull method to constructing the ma-margin plane for our augmented sets + and assuming the are strictl separable, i.e. ɛ > ɛ 0. The problem is illustrated in detail in Figure. The closest points of + and can be found b solving the following dual C-SVR quadratic program: min u,v ( X (+ɛe) ) u ( X ( ɛe) ) v s.t. e u =, e v =, u 0, v 0. (4) Let the closest points in the conve hulls of + and be c = ( )û X (+ɛe) and d = ( X )ˆv ( ɛe) respectivel. The ma-margin separating plane bisects these two points. The normal (ŵ, ˆδ) of the plane is the difference between them, i.e., ŵ = X û X ˆv, ˆδ = ( + ɛe) û ( ɛe) ˆv. The threshold, ˆb, is the distance from the origin to the) point halfwa ) between the two closest points along the normal: ˆb = +ˆδ. The separating plane has the equation ŵ +ˆδ ˆb = 0. ( û+ ˆv ( ŵ X û+x ˆv Rescaling this plane ields the regression function. ual C-SVR (4) is in the dual space. The corresponding Primal C-SVR is:
5 Figure : The solution ˆɛ-tube found b C-SVR can have ˆɛ < ɛ. Squares are original data. ots are in +. Triangles are in. Support Vectors are circled. min w,δ,α,β s.t. w + δ (α β) Xw + δ( + ɛe) αe 0 Xw + δ( ɛe) βe 0. (5) ual C-SVR (4) can be derived b taking the Wolfe or Lagrangian dual [4] of primal C-SVR (5) and simplifing. We prove that the optimal plane from C-SVR bisects the ˆɛ tube. The supporting planes for class + and class determines the lower and upper edges of the ˆɛ-tube respectivel. The support vectors from + and correspond to the points along the lower and upper edges of the ˆɛ-tube. See Figure. Theorem 3. (C-SVR constructs ˆɛ-tube) Let the ma-margin plane obtained b C-SVR ( (4) be ) ŵ +ˆδ ˆb = 0 where ŵ = X û X ˆv, ˆδ = (+ɛe) û ( ɛe) ˆv, and ˆb = ŵ + ˆδ ( ). If ɛ > ɛ 0, then the plane = w + b corresponds X û+x ˆv û+ ˆv to an ˆɛ-tube of training data (X i, i ), i =,,, m where w = ŵˆδ, b = ˆbˆδ ˆɛ = ɛ ˆδ < ɛ. Proof First, we show ˆδ > 0. B the Wolfe dualit theorem [4], ˆα ˆβ > 0, since the objective values of (5) and the negative objective value of (4) are equal at optimalit. B complementarit, the closest points are right on the margin planes ŵ + ˆδ ˆα = 0 and ŵ + ˆδ ˆβ = 0 respectivel, so ˆα = ŵ X û + ˆδ( + ɛe) û and ˆβ = ŵ X ˆv+ˆδ( ɛe)ˆv. Hence ˆb = ˆα+ ˆβ, and ŵ, ˆδ, ˆα, and ˆβ satisf the constraints of problem (5), i.e., Xŵ+ ˆδ(+ɛe) ˆαe 0, Xŵ+ ˆδ( ɛe) ˆβe 0. Then subtract the second inequalit from the first inequalit: ˆδɛ ˆα + ˆβ 0, that is, ˆδ ɛ > 0 because ɛ > ɛ 0 0. Rescale constraints b ˆδ < 0, and reverse the signs. Let w = ŵˆδ, then the inequalities become Xw ɛe ˆαˆδ e, Xw ɛe ˆβ ˆδ e. Let b = ˆbˆδ, then ˆαˆδ = b + and ˆβ = b. Substituting into the previous ˆδ ( ˆδ ) ˆδ ( ) inequalities ields Xw ɛ e be, Xw ɛ e be. enote ˆδ and ˆɛ = ɛ < ɛ. These inequalities become Xw+be ˆɛe, Xw+be ˆɛe. ˆδ Hence the plane = w + b is in the middle of the ˆɛ < ɛ tube. 3. The soft ɛ-tube case For ɛ < ɛ 0, a hard ɛ-tube does not eist. Making ɛ large to fit outliers ma result in poor overall accurac. In soft-margin classification, outliers were handled in the ˆδ
6 ε^ Figure 3: Soft ˆɛ-tube found b RC-SVR: left: dual, right: primal space. dual space b using reduced conve hulls. The same strateg works for soft ɛ-tubes, see Figure 3. Instead of taking the full conve hulls of + and, we reduce the conve hulls awa from the difficult boundar cases. RC-SVR computes the closest points in the reduced conve hulls min u,v s.t. ( ) ( X (+ɛe) u X ) ( ɛe) v e u =, e v =, 0 u e, 0 v e. Parameter determines the robustness of the solution b reducing the conve hull. limits the influence of an single point. As in ν-svm, we can parameterize b ν. Let = where m is the number of points. Figure 3 illustrates the case νm for m = 6 points, ν = /6, and = /. In this eample, ever point in the reduced conve hull must depend on at least two data points since m i= u i = and 0 u i /. In general, ever point in the reduced conve hull can be written as the conve combination of at least / = ν m. Since these points are eactl the support vectors and there are two reduced conve hulls, νm is a lower bound on the number of support vectors in RC-SVR. B choosing ν sufficientl large, the inseparable case with ɛ ɛ 0 is transformed into a separable case where once again our nearest-points-in-the-conve-hull-problem is well defined. As in classification, the dual reduced conve hull problem corresponds to computing a soft ɛ-tube in the primal space. Consider the following soft tube version of the primal C-SVR (7) which has its Wolfe ual RC-SVR (6): min w,δ,α,β,ξ,η w + δ (α β) + C(e ξ + e η) s.t. Xw + δ( + ɛe) αe + ξ 0, ξ 0 Xw + δ( ɛe) βe η 0, η 0 The results of Theorem 3. can be easil etended to soft ɛ-tubes. Theorem 3. (RC-SVR constructs soft ˆɛ-tube) Let the soft ma-margin plane obtained b RC-SVR (6) be ŵ + ˆδ ˆb = 0 where ŵ = X û X ˆv, ˆδ = ( + ɛe) û ( ɛe) ˆv, ( ) ( ) X and ˆb = û+x ˆv ŵ + û+ ˆv ˆδ. If 0 < ɛ ɛ 0, then the plane = w + b corresponds to a soft ˆɛ = ɛ < ɛ-tube of training data ˆδ (X i, i ), i =,,, m, i.e., a ˆɛ-tube of reduced conve hull of training data where w = ŵˆδ, b = ˆbˆδ and α = ŵ X û + ˆδ( + ɛe) û, β = ŵ X ˆv + ˆδ( ɛe) ˆv. Notice that the α and β determine the planes parallel to the regression plane and through the closest points in each reduced conve hull of shifted data. In the α β (6) (7)
7 inseparable case, these planes are parallel but not necessaril identical to the planes obtained b the primal RC-SVR (7). Nonlinear C-SVR and RC-SVR can be achieved b using the usual kernel trick. Let Φ b a nonlinear mapping of such that k(x i, X j ) = Φ(X i ) Φ(X j ). The objective function of C-SVR (4) and RC-SVR (6) applied to the mapped data becomes m m i= j= ((u i v i )(u j v j )(Φ(X i ) Φ(X j ) + i j )) + ɛ m i= ( i(u i v i )) = m m i= j= ((u i v i )(u j v j )(k(x i, X j ) + i j )) + ɛ m i= ( i(u i v i )) (8) The final regression model after optimizing C-SVR or RC-SVR with kernels takes the form of f() = m i= (ū i v i ) k(x i, ) + b, where ū i = ûi ˆδ, v i = ˆvi ˆδ, ˆδ = (û ˆv) +ɛ, and the intercept term b = (û+ˆv) K(û ˆv) + (û+ˆv) ˆδ where K ij = k(x i, X j ). 4 Computational Results We illustrate the difference between RC-SVR and ɛ-svr on a to linear problem 3. Figure 4 depicts the functions constructed b RC-SVR and ɛ-svr for different values of ɛ. For large ɛ, ɛ-svr produces undesirable results. RC-SVR constructs the same function for ɛ sufficientl large. Too small ɛ can result in poor generalization ε = 0.75, 0.45, 0.5 ε = 0.75 ε = ε = ε = (a) (b) Figure 4: Regression lines from (a) ɛ-svr and (b) RC-SVR with distinct ɛ. In Table, we compare RC-SVR, ɛ-svr and ν-svr on the Boston Housing problem. Following the eperimental design in [5] we used RBF kernel with σ = 3.9, C = 500 m for ɛ-svr and ν-svr, and ɛ = 3.0 for RC-SVR. RC-SVR, ɛ-svr, and ν-svr are computationall similar for good parameter choices. In ɛ-svr, ɛ is fied. In RC-SVR, ɛ is the maimum allowable tube width. Choosing ɛ is critical for ɛ-svr but less so for RC-SVR. Both RC-SVR and ν-svr can shrink or grow the tube according to desired robustness. But ν-svr has no upper ɛ bound. 5 Conclusion and iscussion B eamining when ɛ-tubes eist, we showed that in the dual space SVR can be regarded as a classification problem. Hard and soft ɛ-tubes are constructed b separating the conve or reduced conve hulls respectivel of the training data with the response variable shifted up and down b ɛ. We proposed RC-SVR based on choosing the soft ma-margin plane between the two shifted datasets. Like ν-svm, RC-SVR shrinks the ɛ-tube. The ma-margin determines how much the tube can shrink. omain knowledge can be incorporated into the RC-SVR parameters ɛ 3 The data consist of (, ): (0 0), ( 0.), ( 0.7), (.5 0.9), (3.) and (5 ). The CPLEX 6.6 optimization package was used.
8 Table : Testing Results for Boston Housing, MSE= average of mean squared errors of 5 testing points over 00 trials, ST: standard deviation ν RC-SVR MSE ST ɛ ɛ-svr MSE ST ν ν-svr MSE ST and ν. The parameter C in ν-svm and ɛ-svr has been eliminated. Computationall, no one method is superior for good parameter choices. RC-SVR alone has a geometricall intuitive framework that allows users to easil grasp the model and its parameters. Also, RC-SVR can be solved b fast nearest point algorithms. Considering regression as a classification problem suggests other interesting SVR formulations. We can show ɛ-svr is equivalent to finding closest points in a reduced conve hull problem for certain C, but the equivalent problem utilizes a different metric in the objective function than RC-SVR. Perhaps other variations would ield even better formulations. Acknowledgments Thanks to referees and Bernhard Schölkopf for suggestions to improve this work. This work was supported b NSF IRI , NSF IIS References [] K. Bennett and E. Bredensteiner. ualit and Geometr in SVM Classifiers. In P. Langle, eds., Proc. of Seventeenth Intl. Conf. on Machine Learning, p 57 64, Morgan Kaufmann, San Francisco, 000. []. Crisp and C. Burges. A Geometric Interpretation of ν-svm Classifiers. In S. Solla, T. Leen, and K. Muller, eds., Advances in Neural Info. Proc. Ss., Vol. p 44 5, MIT Press, Cambridge, MA, 999. [3] S.S. Keerthi, S.K. Shevade, C. Bhattachara and K.R.K. Murth, A Fast Iterative Nearest Point Algorithm for Support Vector Machine Classifier esign, IEEE Transactions on Neural Networks, Vol., pp.4-36, 000. [4] O. Mangasarian. Nonlinear Programming. SIAM, Philadelphia, 994. [5] B. Schölkopf, P. Bartlett, A. Smola and R. Williamson. Shrinking the Tube: A New Support Vector Regression Algorithm. In M. Kearns, S. Solla, and. Cohn eds., Advances in Neural Info. Proc. Ss., Vol, MIT Press, Cambridge, MA, 999. [6] V. Vapnik. The Nature of Statistical Learning Theor. Wile, New York, 995.
Linear, threshold units. Linear Discriminant Functions and Support Vector Machines. Biometrics CSE 190 Lecture 11. X i : inputs W i : weights
Linear Discriminant Functions and Support Vector Machines Linear, threshold units CSE19, Winter 11 Biometrics CSE 19 Lecture 11 1 X i : inputs W i : weights θ : threshold 3 4 5 1 6 7 Courtesy of University
More informationSupport Vector Machine Classification via Parameterless Robust Linear Programming
Support Vector Machine Classification via Parameterless Robust Linear Programming O. L. Mangasarian Abstract We show that the problem of minimizing the sum of arbitrary-norm real distances to misclassified
More informationA GENERAL FORMULATION FOR SUPPORT VECTOR MACHINES. Wei Chu, S. Sathiya Keerthi, Chong Jin Ong
A GENERAL FORMULATION FOR SUPPORT VECTOR MACHINES Wei Chu, S. Sathiya Keerthi, Chong Jin Ong Control Division, Department of Mechanical Engineering, National University of Singapore 0 Kent Ridge Crescent,
More informationSquare Penalty Support Vector Regression
Square Penalty Support Vector Regression Álvaro Barbero, Jorge López, José R. Dorronsoro Dpto. de Ingeniería Informática and Instituto de Ingeniería del Conocimiento Universidad Autónoma de Madrid, 28049
More informationν =.1 a max. of 10% of training set can be margin errors ν =.8 a max. of 80% of training can be margin errors
p.1/1 ν-svms In the traditional softmargin classification SVM formulation we have a penalty constant C such that 1 C size of margin. Furthermore, there is no a priori guidance as to what C should be set
More informationLearning with Rigorous Support Vector Machines
Learning with Rigorous Support Vector Machines Jinbo Bi 1 and Vladimir N. Vapnik 2 1 Department of Mathematical Sciences, Rensselaer Polytechnic Institute, Troy NY 12180, USA 2 NEC Labs America, Inc. Princeton
More informationPolyhedral Computation. Linear Classifiers & the SVM
Polyhedral Computation Linear Classifiers & the SVM mcuturi@i.kyoto-u.ac.jp Nov 26 2010 1 Statistical Inference Statistical: useful to study random systems... Mutations, environmental changes etc. life
More informationSupport Vector Machines for Classification and Regression. 1 Linearly Separable Data: Hard Margin SVMs
E0 270 Machine Learning Lecture 5 (Jan 22, 203) Support Vector Machines for Classification and Regression Lecturer: Shivani Agarwal Disclaimer: These notes are a brief summary of the topics covered in
More informationIn applications, we encounter many constrained optimization problems. Examples Basis pursuit: exact sparse recovery problem
1 Conve Analsis Main references: Vandenberghe UCLA): EECS236C - Optimiation methods for large scale sstems, http://www.seas.ucla.edu/ vandenbe/ee236c.html Parikh and Bod, Proimal algorithms, slides and
More informationReview: Support vector machines. Machine learning techniques and image analysis
Review: Support vector machines Review: Support vector machines Margin optimization min (w,w 0 ) 1 2 w 2 subject to y i (w 0 + w T x i ) 1 0, i = 1,..., n. Review: Support vector machines Margin optimization
More informationSupport Vector Machines Explained
December 23, 2008 Support Vector Machines Explained Tristan Fletcher www.cs.ucl.ac.uk/staff/t.fletcher/ Introduction This document has been written in an attempt to make the Support Vector Machines (SVM),
More informationSupport Vector Machines
Wien, June, 2010 Paul Hofmarcher, Stefan Theussl, WU Wien Hofmarcher/Theussl SVM 1/21 Linear Separable Separating Hyperplanes Non-Linear Separable Soft-Margin Hyperplanes Hofmarcher/Theussl SVM 2/21 (SVM)
More informationSupport Vector Machines for Classification and Regression
CIS 520: Machine Learning Oct 04, 207 Support Vector Machines for Classification and Regression Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture. They may
More informationRELATIONS AND FUNCTIONS through
RELATIONS AND FUNCTIONS 11.1.2 through 11.1. Relations and Functions establish a correspondence between the input values (usuall ) and the output values (usuall ) according to the particular relation or
More informationQUADRATIC AND CONVEX MINIMAX CLASSIFICATION PROBLEMS
Journal of the Operations Research Societ of Japan 008, Vol. 51, No., 191-01 QUADRATIC AND CONVEX MINIMAX CLASSIFICATION PROBLEMS Tomonari Kitahara Shinji Mizuno Kazuhide Nakata Toko Institute of Technolog
More informationAffine transformations
Reading Optional reading: Affine transformations Brian Curless CSE 557 Autumn 207 Angel and Shreiner: 3., 3.7-3. Marschner and Shirle: 2.3, 2.4.-2.4.4, 6..-6..4, 6.2., 6.3 Further reading: Angel, the rest
More informationAn Optimization Method for Numerically Solving Three-point BVPs of Linear Second-order ODEs with Variable Coefficients
An Optimization Method for Numericall Solving Three-point BVPs of Linear Second-order ODEs with Variable Coefficients Liaocheng Universit School of Mathematics Sciences 252059, Liaocheng P.R. CHINA hougbb@26.com;
More informationKnowledge-Based Nonlinear Kernel Classifiers
Knowledge-Based Nonlinear Kernel Classifiers Glenn M. Fung, Olvi L. Mangasarian, and Jude W. Shavlik Computer Sciences Department, University of Wisconsin Madison, WI 5376 {gfung,olvi,shavlik}@cs.wisc.edu
More informationSupport Vector Machines. CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington
Support Vector Machines CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington 1 A Linearly Separable Problem Consider the binary classification
More informationAn Improved Conjugate Gradient Scheme to the Solution of Least Squares SVM
An Improved Conjugate Gradient Scheme to the Solution of Least Squares SVM Wei Chu Chong Jin Ong chuwei@gatsby.ucl.ac.uk mpeongcj@nus.edu.sg S. Sathiya Keerthi mpessk@nus.edu.sg Control Division, Department
More informationPerceptron Revisited: Linear Separators. Support Vector Machines
Support Vector Machines Perceptron Revisited: Linear Separators Binary classification can be viewed as the task of separating classes in feature space: w T x + b > 0 w T x + b = 0 w T x + b < 0 Department
More informationIntermediate Math Circles Wednesday November Inequalities and Linear Optimization
WWW.CEMC.UWATERLOO.CA The CENTRE for EDUCATION in MATHEMATICS and COMPUTING Intermediate Math Circles Wednesda November 21 2012 Inequalities and Linear Optimization Review: Our goal is to solve sstems
More informationSupport'Vector'Machines. Machine(Learning(Spring(2018 March(5(2018 Kasthuri Kannan
Support'Vector'Machines Machine(Learning(Spring(2018 March(5(2018 Kasthuri Kannan kasthuri.kannan@nyumc.org Overview Support Vector Machines for Classification Linear Discrimination Nonlinear Discrimination
More information12.1 Systems of Linear equations: Substitution and Elimination
. Sstems of Linear equations: Substitution and Elimination Sstems of two linear equations in two variables A sstem of equations is a collection of two or more equations. A solution of a sstem in two variables
More informationConvex Optimization Overview (cnt d)
Conve Optimization Overview (cnt d) Chuong B. Do November 29, 2009 During last week s section, we began our study of conve optimization, the study of mathematical optimization problems of the form, minimize
More informationSupport Vector Machine Regression for Volatile Stock Market Prediction
Support Vector Machine Regression for Volatile Stock Market Prediction Haiqin Yang, Laiwan Chan, and Irwin King Department of Computer Science and Engineering The Chinese University of Hong Kong Shatin,
More informationSupport Vector Regression (SVR) Descriptions of SVR in this discussion follow that in Refs. (2, 6, 7, 8, 9). The literature
Support Vector Regression (SVR) Descriptions of SVR in this discussion follow that in Refs. (2, 6, 7, 8, 9). The literature suggests the design variables should be normalized to a range of [-1,1] or [0,1].
More informationProbabilistic Regression Using Basis Function Models
Probabilistic Regression Using Basis Function Models Gregory Z. Grudic Department of Computer Science University of Colorado, Boulder grudic@cs.colorado.edu Abstract Our goal is to accurately estimate
More informationSupport Vector Machines and Kernel Algorithms
Support Vector Machines and Kernel Algorithms Bernhard Schölkopf Max-Planck-Institut für biologische Kybernetik 72076 Tübingen, Germany Bernhard.Schoelkopf@tuebingen.mpg.de Alex Smola RSISE, Australian
More informationSupport Vector Machine for Classification and Regression
Support Vector Machine for Classification and Regression Ahlame Douzal AMA-LIG, Université Joseph Fourier Master 2R - MOSIG (2013) November 25, 2013 Loss function, Separating Hyperplanes, Canonical Hyperplan
More informationSupport Vector Machine (SVM) and Kernel Methods
Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2014 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin
More informationStrain Transformation and Rosette Gage Theory
Strain Transformation and Rosette Gage Theor It is often desired to measure the full state of strain on the surface of a part, that is to measure not onl the two etensional strains, and, but also the shear
More informationSupport Vector Machine (SVM) & Kernel CE-717: Machine Learning Sharif University of Technology. M. Soleymani Fall 2012
Support Vector Machine (SVM) & Kernel CE-717: Machine Learning Sharif University of Technology M. Soleymani Fall 2012 Linear classifier Which classifier? x 2 x 1 2 Linear classifier Margin concept x 2
More information14.1 Systems of Linear Equations in Two Variables
86 Chapter 1 Sstems of Equations and Matrices 1.1 Sstems of Linear Equations in Two Variables Use the method of substitution to solve sstems of equations in two variables. Use the method of elimination
More informationImprovements to Platt s SMO Algorithm for SVM Classifier Design
LETTER Communicated by John Platt Improvements to Platt s SMO Algorithm for SVM Classifier Design S. S. Keerthi Department of Mechanical and Production Engineering, National University of Singapore, Singapore-119260
More informationA set C R n is convex if and only if δ C is convex if and only if. f : R n R is strictly convex if and only if f is convex and the inequality (1.
ONVEX OPTIMIZATION SUMMARY ANDREW TULLOH 1. Eistence Definition. For R n, define δ as 0 δ () = / (1.1) Note minimizes f over if and onl if minimizes f + δ over R n. Definition. (ii) (i) dom f = R n f()
More informationIntroduction to Support Vector Machines
Introduction to Support Vector Machines Andreas Maletti Technische Universität Dresden Fakultät Informatik June 15, 2006 1 The Problem 2 The Basics 3 The Proposed Solution Learning by Machines Learning
More informationCh 3 Alg 2 Note Sheet.doc 3.1 Graphing Systems of Equations
Ch 3 Alg Note Sheet.doc 3.1 Graphing Sstems of Equations Sstems of Linear Equations A sstem of equations is a set of two or more equations that use the same variables. If the graph of each equation =.4
More informationLECTURE 7 Support vector machines
LECTURE 7 Support vector machines SVMs have been used in a multitude of applications and are one of the most popular machine learning algorithms. We will derive the SVM algorithm from two perspectives:
More informationAn introduction to Support Vector Machines
1 An introduction to Support Vector Machines Giorgio Valentini DSI - Dipartimento di Scienze dell Informazione Università degli Studi di Milano e-mail: valenti@dsi.unimi.it 2 Outline Linear classifiers
More informationNeural Networks. Prof. Dr. Rudolf Kruse. Computational Intelligence Group Faculty for Computer Science
Neural Networks Prof. Dr. Rudolf Kruse Computational Intelligence Group Faculty for Computer Science kruse@iws.cs.uni-magdeburg.de Rudolf Kruse Neural Networks 1 Supervised Learning / Support Vector Machines
More informationSMO Algorithms for Support Vector Machines without Bias Term
Institute of Automatic Control Laboratory for Control Systems and Process Automation Prof. Dr.-Ing. Dr. h. c. Rolf Isermann SMO Algorithms for Support Vector Machines without Bias Term Michael Vogt, 18-Jul-2002
More informationOutline. Basic concepts: SVM and kernels SVM primal/dual problems. Chih-Jen Lin (National Taiwan Univ.) 1 / 22
Outline Basic concepts: SVM and kernels SVM primal/dual problems Chih-Jen Lin (National Taiwan Univ.) 1 / 22 Outline Basic concepts: SVM and kernels Basic concepts: SVM and kernels SVM primal/dual problems
More informationSupport Vector Machine (continued)
Support Vector Machine continued) Overlapping class distribution: In practice the class-conditional distributions may overlap, so that the training data points are no longer linearly separable. We need
More informationStatistical Pattern Recognition
Statistical Pattern Recognition Support Vector Machine (SVM) Hamid R. Rabiee Hadi Asheri, Jafar Muhammadi, Nima Pourdamghani Spring 2013 http://ce.sharif.edu/courses/91-92/2/ce725-1/ Agenda Introduction
More informationSupport Vector Machine (SVM) and Kernel Methods
Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2015 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin
More informationSupport Vector Machines. Maximizing the Margin
Support Vector Machines Support vector achines (SVMs) learn a hypothesis: h(x) = b + Σ i= y i α i k(x, x i ) (x, y ),..., (x, y ) are the training exs., y i {, } b is the bias weight. α,..., α are the
More informationMinOver Revisited for Incremental Support-Vector-Classification
MinOver Revisited for Incremental Support-Vector-Classification Thomas Martinetz Institute for Neuro- and Bioinformatics University of Lübeck D-23538 Lübeck, Germany martinetz@informatik.uni-luebeck.de
More informationSVM Incremental Learning, Adaptation and Optimization
SVM Incremental Learning, Adaptation and Optimization Christopher P. Diehl Applied Physics Laboratory Johns Hopkins University Laurel, MD 20723 Chris.Diehl@jhuapl.edu Gert Cauwenberghs ECE Department Johns
More informationSupport Vector Machine (SVM) and Kernel Methods
Support Vector Machine (SVM) and Kernel Methods CE-717: Machine Learning Sharif University of Technology Fall 2016 Soleymani Outline Margin concept Hard-Margin SVM Soft-Margin SVM Dual Problems of Hard-Margin
More informationIntroduction to SVM and RVM
Introduction to SVM and RVM Machine Learning Seminar HUS HVL UIB Yushu Li, UIB Overview Support vector machine SVM First introduced by Vapnik, et al. 1992 Several literature and wide applications Relevance
More informationSUPPORT VECTOR MACHINE FOR THE SIMULTANEOUS APPROXIMATION OF A FUNCTION AND ITS DERIVATIVE
SUPPORT VECTOR MACHINE FOR THE SIMULTANEOUS APPROXIMATION OF A FUNCTION AND ITS DERIVATIVE M. Lázaro 1, I. Santamaría 2, F. Pérez-Cruz 1, A. Artés-Rodríguez 1 1 Departamento de Teoría de la Señal y Comunicaciones
More informationDiscriminative Direction for Kernel Classifiers
Discriminative Direction for Kernel Classifiers Polina Golland Artificial Intelligence Lab Massachusetts Institute of Technology Cambridge, MA 02139 polina@ai.mit.edu Abstract In many scientific and engineering
More informationFoundation of Intelligent Systems, Part I. SVM s & Kernel Methods
Foundation of Intelligent Systems, Part I SVM s & Kernel Methods mcuturi@i.kyoto-u.ac.jp FIS - 2013 1 Support Vector Machines The linearly-separable case FIS - 2013 2 A criterion to select a linear classifier:
More informationLinear programming: Theory
Division of the Humanities and Social Sciences Ec 181 KC Border Convex Analsis and Economic Theor Winter 2018 Topic 28: Linear programming: Theor 28.1 The saddlepoint theorem for linear programming The
More informationA Revisit to Support Vector Data Description (SVDD)
A Revisit to Support Vector Data Description (SVDD Wei-Cheng Chang b99902019@csie.ntu.edu.tw Ching-Pei Lee r00922098@csie.ntu.edu.tw Chih-Jen Lin cjlin@csie.ntu.edu.tw Department of Computer Science, National
More informationStatistical Machine Learning from Data
Samy Bengio Statistical Machine Learning from Data 1 Statistical Machine Learning from Data Support Vector Machines Samy Bengio IDIAP Research Institute, Martigny, Switzerland, and Ecole Polytechnique
More informationKernel Methods and Support Vector Machines
Kernel Methods and Support Vector Machines Oliver Schulte - CMPT 726 Bishop PRML Ch. 6 Support Vector Machines Defining Characteristics Like logistic regression, good for continuous input features, discrete
More informationRobust Kernel-Based Regression
Robust Kernel-Based Regression Budi Santosa Department of Industrial Engineering Sepuluh Nopember Institute of Technology Kampus ITS Surabaya Surabaya 60111,Indonesia Theodore B. Trafalis School of Industrial
More informationLarge Scale Kernel Regression via Linear Programming
Large Scale Kernel Regression via Linear Programming O. L. Mangasarian David R. Musicant Computer Sciences Dept. Dept. of Mathematics and Computer Science University of Wisconsin Carleton College 20 West
More informationLinear & nonlinear classifiers
Linear & nonlinear classifiers Machine Learning Hamid Beigy Sharif University of Technology Fall 1394 Hamid Beigy (Sharif University of Technology) Linear & nonlinear classifiers Fall 1394 1 / 34 Table
More informationKernel PCA, clustering and canonical correlation analysis
ernel PCA, clustering and canonical correlation analsis Le Song Machine Learning II: Advanced opics CSE 8803ML, Spring 2012 Support Vector Machines (SVM) 1 min w 2 w w + C j ξ j s. t. w j + b j 1 ξ j,
More informationSample-based Optimal Transport and Barycenter Problems
Sample-based Optimal Transport and Barcenter Problems MAX UANG New York Universit, Courant Institute of Mathematical Sciences AND ESTEBAN G. TABA New York Universit, Courant Institute of Mathematical Sciences
More informationGlossary. Also available at BigIdeasMath.com: multi-language glossary vocabulary flash cards. An equation that contains an absolute value expression
Glossar This student friendl glossar is designed to be a reference for ke vocabular, properties, and mathematical terms. Several of the entries include a short eample to aid our understanding of important
More informationReview: critical point or equivalently f a,
Review: a b f f a b f a b critical point or equivalentl f a, b A point, is called a of if,, 0 A local ma or local min must be a critical point (but not conversel) 0 D iscriminant (or Hessian) f f D f f
More informationTHE ADVICEPTRON: GIVING ADVICE TO THE PERCEPTRON
THE ADVICEPTRON: GIVING ADVICE TO THE PERCEPTRON Gautam Kunapuli Kristin P. Bennett University of Wisconsin Madison Rensselaer Polytechnic Institute Madison, WI 5375 Troy, NY 1218 kunapg@wisc.edu bennek@rpi.edu
More informationForecast daily indices of solar activity, F10.7, using support vector regression method
Research in Astron. Astrophys. 9 Vol. 9 No. 6, 694 702 http://www.raa-journal.org http://www.iop.org/journals/raa Research in Astronomy and Astrophysics Forecast daily indices of solar activity, F10.7,
More informationChapter Adequacy of Solutions
Chapter 04.09 dequac of Solutions fter reading this chapter, ou should be able to: 1. know the difference between ill-conditioned and well-conditioned sstems of equations,. define the norm of a matri,
More informationSupport Vector Machines. Machine Learning Fall 2017
Support Vector Machines Machine Learning Fall 2017 1 Where are we? Learning algorithms Decision Trees Perceptron AdaBoost 2 Where are we? Learning algorithms Decision Trees Perceptron AdaBoost Produce
More informationSupport Vector Machines
Support Vector Machines Support vector machines (SVMs) are one of the central concepts in all of machine learning. They are simply a combination of two ideas: linear classification via maximum (or optimal
More informationNON-FIXED AND ASYMMETRICAL MARGIN APPROACH TO STOCK MARKET PREDICTION USING SUPPORT VECTOR REGRESSION. Haiqin Yang, Irwin King and Laiwan Chan
In The Proceedings of ICONIP 2002, Singapore, 2002. NON-FIXED AND ASYMMETRICAL MARGIN APPROACH TO STOCK MARKET PREDICTION USING SUPPORT VECTOR REGRESSION Haiqin Yang, Irwin King and Laiwan Chan Department
More informationEigenvectors and Eigenvalues 1
Ma 2015 page 1 Eigenvectors and Eigenvalues 1 In this handout, we will eplore eigenvectors and eigenvalues. We will begin with an eploration, then provide some direct eplanation and worked eamples, and
More informationPattern Recognition 2018 Support Vector Machines
Pattern Recognition 2018 Support Vector Machines Ad Feelders Universiteit Utrecht Ad Feelders ( Universiteit Utrecht ) Pattern Recognition 1 / 48 Support Vector Machines Ad Feelders ( Universiteit Utrecht
More informationEquations of lines in
Roberto s Notes on Linear Algebra Chapter 6: Lines, planes an other straight objects Section 1 Equations of lines in What ou nee to know alrea: The ot prouct. The corresponence between equations an graphs.
More informationSupport Vector Machines. Introduction to Data Mining, 2 nd Edition by Tan, Steinbach, Karpatne, Kumar
Data Mining Support Vector Machines Introduction to Data Mining, 2 nd Edition by Tan, Steinbach, Karpatne, Kumar 02/03/2018 Introduction to Data Mining 1 Support Vector Machines Find a linear hyperplane
More informationMachine Learning. Kernels. Fall (Kernels, Kernelized Perceptron and SVM) Professor Liang Huang. (Chap. 12 of CIML)
Machine Learning Fall 2017 Kernels (Kernels, Kernelized Perceptron and SVM) Professor Liang Huang (Chap. 12 of CIML) Nonlinear Features x4: -1 x1: +1 x3: +1 x2: -1 Concatenated (combined) features XOR:
More informationMachine Learning. Support Vector Machines. Fabio Vandin November 20, 2017
Machine Learning Support Vector Machines Fabio Vandin November 20, 2017 1 Classification and Margin Consider a classification problem with two classes: instance set X = R d label set Y = { 1, 1}. Training
More informationA Finite Newton Method for Classification Problems
A Finite Newton Method for Classification Problems O. L. Mangasarian Computer Sciences Department University of Wisconsin 1210 West Dayton Street Madison, WI 53706 olvi@cs.wisc.edu Abstract A fundamental
More informationHOMEWORK 4: SVMS AND KERNELS
HOMEWORK 4: SVMS AND KERNELS CMU 060: MACHINE LEARNING (FALL 206) OUT: Sep. 26, 206 DUE: 5:30 pm, Oct. 05, 206 TAs: Simon Shaolei Du, Tianshu Ren, Hsiao-Yu Fish Tung Instructions Homework Submission: Submit
More informationIndirect Rule Learning: Support Vector Machines. Donglin Zeng, Department of Biostatistics, University of North Carolina
Indirect Rule Learning: Support Vector Machines Indirect learning: loss optimization It doesn t estimate the prediction rule f (x) directly, since most loss functions do not have explicit optimizers. Indirection
More informationChapter 5: Quadratic Equations and Functions 5.1 Modeling Data With Quadratic Functions Quadratic Functions and Their Graphs
Ch 5 Alg Note Sheet Ke Chapter 5: Quadratic Equations and Functions 5.1 Modeling Data With Quadratic Functions Quadratic Functions and Their Graphs Definition: Standard Form of a Quadratic Function The
More informationLecture Support Vector Machine (SVM) Classifiers
Introduction to Machine Learning Lecturer: Amir Globerson Lecture 6 Fall Semester Scribe: Yishay Mansour 6.1 Support Vector Machine (SVM) Classifiers Classification is one of the most important tasks in
More informationOverview. IAML: Linear Regression. Examples of regression problems. The Regression Problem
3 / 38 Overview 4 / 38 IAML: Linear Regression Nigel Goddard School of Informatics Semester The linear model Fitting the linear model to data Probabilistic interpretation of the error function Eamples
More informationSupport Vector Machines
Support Vector Machines Tobias Pohlen Selected Topics in Human Language Technology and Pattern Recognition February 10, 2014 Human Language Technology and Pattern Recognition Lehrstuhl für Informatik 6
More informationMachine Learning. Support Vector Machines. Manfred Huber
Machine Learning Support Vector Machines Manfred Huber 2015 1 Support Vector Machines Both logistic regression and linear discriminant analysis learn a linear discriminant function to separate the data
More informationKernel Methods and Support Vector Machines
Kernel Methods and Support Vector Machines Bernhard Schölkopf Max-Planck-Institut für biologische Kybernetik 72076 Tübingen, Germany Bernhard.Schoelkopf@tuebingen.mpg.de Alex Smola RSISE, Australian National
More informationLecture 18: Kernels Risk and Loss Support Vector Regression. Aykut Erdem December 2016 Hacettepe University
Lecture 18: Kernels Risk and Loss Support Vector Regression Aykut Erdem December 2016 Hacettepe University Administrative We will have a make-up lecture on next Saturday December 24, 2016 Presentations
More informationIntroduction to Machine Learning Spring 2018 Note Duality. 1.1 Primal and Dual Problem
CS 189 Introduction to Machine Learning Spring 2018 Note 22 1 Duality As we have seen in our discussion of kernels, ridge regression can be viewed in two ways: (1) an optimization problem over the weights
More informationCheng Soon Ong & Christian Walder. Canberra February June 2018
Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 Outlines Overview Introduction Linear Algebra Probability Linear Regression
More informationLinear Programming. Maximize the function. P = Ax + By + C. subject to the constraints. a 1 x + b 1 y < c 1 a 2 x + b 2 y < c 2
Linear Programming Man real world problems require the optimization of some function subject to a collection of constraints. Note: Think of optimizing as maimizing or minimizing for MATH1010. For eample,
More informationSystems of Linear Equations: Solving by Graphing
8.1 Sstems of Linear Equations: Solving b Graphing 8.1 OBJECTIVE 1. Find the solution(s) for a set of linear equations b graphing NOTE There is no other ordered pair that satisfies both equations. From
More informationEE613 Machine Learning for Engineers. Kernel methods Support Vector Machines. jean-marc odobez 2015
EE613 Machine Learning for Engineers Kernel methods Support Vector Machines jean-marc odobez 2015 overview Kernel methods introductions and main elements defining kernels Kernelization of k-nn, K-Means,
More informationData Mining. Linear & nonlinear classifiers. Hamid Beigy. Sharif University of Technology. Fall 1396
Data Mining Linear & nonlinear classifiers Hamid Beigy Sharif University of Technology Fall 1396 Hamid Beigy (Sharif University of Technology) Data Mining Fall 1396 1 / 31 Table of contents 1 Introduction
More informationSupport Vector Machines II. CAP 5610: Machine Learning Instructor: Guo-Jun QI
Support Vector Machines II CAP 5610: Machine Learning Instructor: Guo-Jun QI 1 Outline Linear SVM hard margin Linear SVM soft margin Non-linear SVM Application Linear Support Vector Machine An optimization
More informationSupport Vector Learning for Ordinal Regression
Support Vector Learning for Ordinal Regression Ralf Herbrich, Thore Graepel, Klaus Obermayer Technical University of Berlin Department of Computer Science Franklinstr. 28/29 10587 Berlin ralfh graepel2
More informationChapter 9. Support Vector Machine. Yongdai Kim Seoul National University
Chapter 9. Support Vector Machine Yongdai Kim Seoul National University 1. Introduction Support Vector Machine (SVM) is a classification method developed by Vapnik (1996). It is thought that SVM improved
More informationIntroduction to Support Vector Machines
Introduction to Support Vector Machines Hsuan-Tien Lin Learning Systems Group, California Institute of Technology Talk in NTU EE/CS Speech Lab, November 16, 2005 H.-T. Lin (Learning Systems Group) Introduction
More informationFlows and Connectivity
Chapter 4 Flows and Connectivit 4. Network Flows Capacit Networks A network N is a connected weighted loopless graph (G,w) with two specified vertices and, called the source and the sink, respectivel,
More informationYou don't have to be a mathematician to have a feel for numbers. John Forbes Nash, Jr.
Course Title: Real Analsis Course Code: MTH3 Course instructor: Dr. Atiq ur Rehman Class: MSc-II Course URL: www.mathcit.org/atiq/fa5-mth3 You don't have to be a mathematician to have a feel for numbers.
More informationSupport Vector Regression with Automatic Accuracy Control B. Scholkopf y, P. Bartlett, A. Smola y,r.williamson FEIT/RSISE, Australian National University, Canberra, Australia y GMD FIRST, Rudower Chaussee
More information