Generalized Models Based on Neural Networks and Multiple Linear Regression

Size: px
Start display at page:

Download "Generalized Models Based on Neural Networks and Multiple Linear Regression"

Transcription

1 Proceedings of the 5th WSEAS Int. Conf. on CIRCUITS, SYSTEMS, ELECTRONICS, CONTROL & SIGNAL PROCESSING, Dallas, USA, November 1-3, Generalized Models Based on Neural Networks and Multiple Linear Regression PERO RADONJA 1, SRDJAN STANKOVIC 2, DRAGANA DRAZIC 1 and BRATISLAV MATOVIC 1 1 Institute of Forestry, 113 Kneza Viseslava 3, Belgrade 2 Faculty of Electrical Engineering, University of Belgrade, Belgrade SERBIA radonjap@eunet.yu stankovic@etf.bg.ac.yu inszasum@eunet.yu Abstract - Developed generalized models are based on neural networks, linear and multiple linear regression. Applications of the generalized (regional) models in estimation of the most important general (common) characteristic of the whole region is done. Testing of the regional models with real, referent data is performed by regression analyses, also. The obtained correlation coefficients between referent data and corresponding data computed using the regional models based on linear and multiple linear regression are very height,.9971 and.9955 respectively. In the case of application of neural networks correlation coefficients is the largest, Key Words: - generalized model, regional model, multiple linear regression, neural networks, Vapnik- Chervonenkis dimension 1 Introduction Regional (generalized) process models can be obtained by introduction some limits in the parameters of general process models. The process which will be considered is biological process of growing. A region can contain many thousands of individual objects. To get the information concerning with same region, it is necessary to make measurements in several (r) samples with more (N) objects. The total number of objects in all samples, rxn, must encompass the complete variability of the region. In our case there were 156 objects in all samples. Detailed measurements on all objects is expensive and requires a lot of time. Because of this we try to find the regional model that enables to get important characteristics of entire region without detailed measurements on all objects. It is known that neural networks (NN) ensure smaller modeling error than classical methods and because of that they are very suitable for generating both specifics and regional models, [1, 2, 3]. *Research is supported by the Ministry of Science and Environmental Protection of Serbia, as part of the project: EE B 2 Problem Definition The classical approach to modeling of this process is based on the modified Brink s function [4]. The canonical form of this function is: px qx y( x) = u+ ve we...(1) Note that the y(x), (1), very good represents considered process, [4, 5, 6]. Parameters u, v and w are function of the basic measured values, D and H, and parameters i, p and q. For every object in the region, [7, 8], it is possible to determine its specific model, that is, parameters i, p and q, by some optimization procedure, [9]. The optimization procedure requires besides the basic measured values D and H, minimum 3 sets of new measured data. In our case we have used 14 data sets for every object. Main goal is to get regional models so that by application of regional models we can get any specific model or important regional characteristic only using the basic measured values D and H. For example: the important regional characteristic is the sum of the volumes of all objects. The volumes of objects from whole region are important because of sustainable management with considered ecological system [1, 11, 12 ]. 3 Specific Models Based on

2 Proceedings of the 5th WSEAS Int. Conf. on CIRCUITS, SYSTEMS, ELECTRONICS, CONTROL & SIGNAL PROCESSING, Dallas, USA, November 1-3, Neural Networks In considered case NN ensure smaller modeling error than classical methods based on polynomial, exponential functions, or even on the modified Brink s function. Because of that, in our case, NN are very suitable tools in process of generating the specific and after that, the generalized model. It is known that to achieve good performance, both the bias and the variance of the model, should be small. In order to avoid overfitting and to get a small confidence interval it is necessary to use NN with low Vapnik-Chervonenkis (VC) dimension [1, 2, 3]. In our cases VC dimension is determined simply by the number of free model parameters, that is by number of tansig neurons in the second (hidden) layer. tansig neurons have logistic sigmoid tangent hyperbolic transfer function. Note that a decrease of VC dimension occurs at the expense of an increase of the training error. The method of structural risk minimization (SRM) provides a systematic procedure for achieving the best compromise between the training error and the confidence interval by controlling the VC dimension [1, 2]. The goal is to find a network structure such that a decrease of the VC dimension occurs at the expense of the smallest possible increase in the training error. In our case, we decrease monotonically the number of tansig neurons in the hidden layers, [13]. The process of learning is shown in Fig.1. In the process of learning, Levenberg-Marqurdt algorithm is used because of its good convergence properties, [14, 15, 16, 17]. Fig.2 Normalized specific model The normalization is performed evidently by using the largest value of x and y coordinates. Errors of modeling is shown in Fig.3 Fig.3 Errors of modeling 4 Generalized Model Obtained by Using NN The generalized model for considered region, that is, the regional model is obtained as mean value of all normalized specific models from all samples. The obtained model, with normalized y and x coordinates, is presented in Fig y Fig.1 Learning process of NN In Fig.2, the normalized specific model is presented x Fig.4 The generalized (regional) model

3 Proceedings of the 5th WSEAS Int. Conf. on CIRCUITS, SYSTEMS, ELECTRONICS, CONTROL & SIGNAL PROCESSING, Dallas, USA, November 1-3, Now any specific model can be obtained by renormalization with D and H, y and x coordinates of the presented generalized model in Fig.4. Quality and performances of the obtained generalized model will be analyzed in the subsequent discussion S = r = Generalized Model Based on Linear Regression Alternatively to Fig.4, the regional model, can be the set equations that determine parameters i, p and q using the basic measured values D and H. In order to analyze the most important characteristics of the considered process and get the regional model, in the first step, the parameters i, p and q for all sobjects are found by some optimization procedure, [9]. In Fig.5 parameter i versus H is presented S = r = i D[cm] Fig.6 Parameter i versus D Parameter q versus D is presented in Fig.7. i S =.4281 r = q H [m].15 Fig.5 Parameter i versus H.1 It can be seen that large correlation,.9541, exist between parameter i and H. That means it is possible to compute i only based on H. In the next figure, Fig.6, as an additional illustration, parameter i versus D is presented.. Now, it can be seen that, the larger correlation coefficient, r,.9976, exist between i and D. Because of this, the parameter i will be computed based on Fig.6, that is on eq.(2). i= D...(2) D[cm] Fig.7 Parameter q versus D It can be seen that correlation coefficient is only Unfortunately, the correlation coefficient, r, with respect to H, is even less,.398. Because of that it will be used the next equation:

4 Proceedings of the 5th WSEAS Int. Conf. on CIRCUITS, SYSTEMS, ELECTRONICS, CONTROL & SIGNAL PROCESSING, Dallas, USA, November 1-3, q= d...(3) Finally, will be analyzed regression lines for parameter p versus D and H. Like in previous case, r of regression process with respect to H, 4336, is less than with respect to D, Because of that, parameter p versus D is presented in Fig.8 opposite to.9541 and.9976, Figs.5 and 6, in the case of simple linear regression. Standard error of modeling, S, is also less,.4175, than and.4256 in Figs.5 and 6. As an illustration of the multiple linear regression, eq.(5), parameter i versus the D and H is presented on Fig.A1, Appendix A. p S = r = Multiple regression offers better results with respect to r, in the case of estimating q parameter, also. r has value.4348 compared with.398 and.4328 in the case simple linear regression, with respect to H an D respectively. Parameter q versus the D and H is presented on Fig D[cm] Fig.8 Parameter p versus D q D 4 2 H Also, it is better to use quadratic fitting than linear one and consequently for computing (estimating) parameter p will be used: p D D 2 = (4) 6 Multiple Linear Regression in Model Generating Generally, multiple regression offers greater possibilities than simple linear regression. With reference to this fact the analysis of applications multiple linear regression is performed also. Based on available data, using multiple linear regression, three equations, for i, q and p parameters, are obtained. i = D H...(5) q = D +.65H...(6) p = D +.865H...(7) In the case of compute (estimation) parameter i by eq.(5) equivalent correlation coefficient is.9977 Fig.9 Parameter q versus the D and H In the case of parameter p estimation by eq.(7), equivalent correlation coefficient is.451 opposite to,5129 and.4336, in the case of the simple linear regression. Illustration of an application of the multiple linear regression for estimation of parameter p, eq.(7), is presented on Fig.A2, Appendix B. 7 Testing the Obtained Generalized (Regional) Models Usability of the regional models will be tasted by computing of the correlation coefficients, that is, by applications of regression analyses technique (method). The very important general (common) characteristics of the region are the volumes of all objects in region and sum of all volumes. The volume of every object depends on the basic measured value D and H, and regional model. The models can be, once obtained by NN (Fig.4) or based on linear regression eq.(2, 3 and 4) or multiple linear regression, eq.(5, 6 and 7).

5 Proceedings of the 5th WSEAS Int. Conf. on CIRCUITS, SYSTEMS, ELECTRONICS, CONTROL & SIGNAL PROCESSING, Dallas, USA, November 1-3, Note, that it is possible to compute real (referent) volumes, Vref, of all objects from considered samples, using the known parameters i, p and q, Figs.5 to 8., [4, 9]. On the other side, we can compute the volumes for the same objects using the regional models, volumes Vm. Testing the accuracy of the regional models is practically performed by computing the correlation coefficients between the referent volumes, Vref, and the volumes Vm. The result of testing, in the case of application of the linear regression, is presented on Fig V m [m 3 ] S =.6744 r = V ref [m 3 ] Fig.1 Testing of application of linear regression based only on values of D Obviously, in the ideal case, the regression line in Fig.1 must have the angle of 45 and has to start from the origin (,) In Fig.11, the result of testing, of model based on multiple linear regression, is presented. Unexpected, the application of multiple linear regression give the worse result, tangα=.9591 and start point.317 where the application of the simple linear regression gives.9678 and.213. Also, value of S and r are worse in the case of the multiple linear regression,.827 and.9955 compared with.674 and.9971 in the case of the simple linear regression V m [m 3 ] S =.827 r = V ref [m 3 ] Fig.11 Testing of regional model based on multiple linear regression Testing the accuracy of the regional models based on NN shows its advantage compared even with model based on the linear regression. Indeed, S and r have value.371 and.9985 compared with.674 and However, tangα has value.9454 that is less worse than.9678 but better start point.53, compared with Conclusion It is shown that the obtained regional models can be very successfully used in the process of computing (estimating) the most important regional characteristics. The volumes of all objects based on regional models are testing with referent volumes and very height correlation coefficients are obtained. Also, it is shown that applications of NN ensures the regional model better than those obtained by the linear and multiple linear regression. References: [1]Haykin, S. (1994): Neural Networks: A Comprehensive Foundation, Macmillan College Publishing Company, New York, [2]Vapnik, Vladimir N. (1999): An Overview of Statistical Learning Theory, IEEE Transaction on Neural Networks, Vol.1, No.5, Sept [3]Radonja Pero, Srdjan Stankovic i Zoran Popovic (24): Specific Process Models Obtained From General Ones, WSEAS TRANSACTIONS on SYSTEMS, Issue 7, Volume 3, September 24, Pub. by WSEAS Press, Athens, pp

6 Proceedings of the 5th WSEAS Int. Conf. on CIRCUITS, SYSTEMS, ELECTRONICS, CONTROL & SIGNAL PROCESSING, Dallas, USA, November 1-3, [4]Riemer, T., von Gadow, K., Sloboda., B. (1995): Ein Modell zur Beschreibung von Baumschaften, Allg Forst Jagdztg 166(7): [5]Hui,G.Y.,von Gadow, K. (1997) Entwicklung und Еrprobung eines Еinheitsschaftmodells fuеr die Baumart Cunninghamia lanceolata. Forstw Cbl 116: [6]Korol M.,von Gadow, K. (23):Ein Einheitsschaftmodell fuer die Baumart Fichte. Forstw Cbl 122: 1-8. [7] Matović, B. (25): Normal state in spruce-fir forest goals and problems in management in Zlatar. MSc thesis, Faculty of forestry, Belgrade. (in Serbian). [8] Maunaga Z. (1995): Productivity and structural characeristics of same-age stands of spruce in Republic Serbian, Ph. thesis, Faculty of forestry, Belgrade, (in Serbian). [9] Radonja, P., Koprivica, M., Matovic, B. (25): Modelling stem profile and volume by using the modified Brink s function, Forestry, N 4, Vol. (LVII):1-1, (in Serbian). [1] Lek, S., A. Belaud, P. Baran, I. Dimopoulos and M. Delacoste, Application of neural networks to modelling nonlinear relationships in ecology, Ecol. Model., 1996, vol. 9, pp [11] Roadknight, C. M., G.R. Balls, G. E. Mills, and D. Palmer-Brown, Modeling complex environmental data, IEEE Trans. Neural Net.,Vol , pp [12] Georgina Stegmayer, Marko Pirola, Giancarlo Orengo and Omar Chiotti: Towards a Volterra series representation from a Neural Network model, Proceedings of the 5th WSEAS Int. Conf. on Neural Networks and Applications, WSEAS NNA 24, Udine, Italy, March 25-27, 24 [13]Beale, Mark, (1993): Neural Network Toolbox NN Toolbox, Version , Release12, September 22, 2, MATLAB 6 R12. [14]Radonja, J.P. (2) Radial Basis Function Neural Networks in Tracking and Extraction of Stochastic Process in Forestry, Proceedings of the 5th Seminar on Neural Networks Application in Electrical Engineering, NEUREL2, September 25-27, 2, IEEE and Acad. Mind, Belgrade, Yug., pp [15]Radonja, J.P., Stankovic S. S. and Cukanovic Nj. R. (2) Multilayer neural networks in process of height curve fitting, INFO science 3/2, Savpo, Belgrade, pp [16]Radonja, Pero, and Stankovivc Srdjan (22): Modeling of a highly nonlinear stochastic process by neural Networks, Recent Advances in Computers, Computing and Communications Editors: N. Mastorakis and V. Mladenov, Published by WSEAS Press. [17] Radonja, Pero, and Stankovic Srdjan: Neural Network Models Based on Small Data Sets, Proceedings of the 6th Seminar on Neural Networks Application in Electrical Engineering, NEUREL22, Editors: B. Reljin and S. Stankovic, September 25-27, Published by: Academic Mind and Faculty of Electrical Engineering, Belgrade, 22, pp Appendix A i D 4 6 Fig.A1 Parameter i versus the D and H Appendix B p D H Fig.A2 Parameter p versus the D and H 4 2 H

Towards a Volterra series representation from a Neural Network model

Towards a Volterra series representation from a Neural Network model Towards a Volterra series representation from a Neural Network model GEORGINA STEGMAYER, MARCO PIROLA Electronics Department Politecnico di Torino Cso. Duca degli Abruzzi 4 9 Torino ITALY GIANCARLO ORENGO

More information

CSC321 Lecture 5: Multilayer Perceptrons

CSC321 Lecture 5: Multilayer Perceptrons CSC321 Lecture 5: Multilayer Perceptrons Roger Grosse Roger Grosse CSC321 Lecture 5: Multilayer Perceptrons 1 / 21 Overview Recall the simple neuron-like unit: y output output bias i'th weight w 1 w2 w3

More information

Support Vector Regression (SVR) Descriptions of SVR in this discussion follow that in Refs. (2, 6, 7, 8, 9). The literature

Support Vector Regression (SVR) Descriptions of SVR in this discussion follow that in Refs. (2, 6, 7, 8, 9). The literature Support Vector Regression (SVR) Descriptions of SVR in this discussion follow that in Refs. (2, 6, 7, 8, 9). The literature suggests the design variables should be normalized to a range of [-1,1] or [0,1].

More information

A Feature Based Neural Network Model for Weather Forecasting

A Feature Based Neural Network Model for Weather Forecasting World Academy of Science, Engineering and Technology 4 2 A Feature Based Neural Network Model for Weather Forecasting Paras, Sanjay Mathur, Avinash Kumar, and Mahesh Chandra Abstract Weather forecasting

More information

epochs epochs

epochs epochs Neural Network Experiments To illustrate practical techniques, I chose to use the glass dataset. This dataset has 214 examples and 6 classes. Here are 4 examples from the original dataset. The last values

More information

Understanding Generalization Error: Bounds and Decompositions

Understanding Generalization Error: Bounds and Decompositions CIS 520: Machine Learning Spring 2018: Lecture 11 Understanding Generalization Error: Bounds and Decompositions Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the

More information

CSC 411 Lecture 10: Neural Networks

CSC 411 Lecture 10: Neural Networks CSC 411 Lecture 10: Neural Networks Roger Grosse, Amir-massoud Farahmand, and Juan Carrasquilla University of Toronto UofT CSC 411: 10-Neural Networks 1 / 35 Inspiration: The Brain Our brain has 10 11

More information

Machine Learning (CSE 446): Neural Networks

Machine Learning (CSE 446): Neural Networks Machine Learning (CSE 446): Neural Networks Noah Smith c 2017 University of Washington nasmith@cs.washington.edu November 6, 2017 1 / 22 Admin No Wednesday office hours for Noah; no lecture Friday. 2 /

More information

Machine Learning

Machine Learning Machine Learning 10-601 Tom M. Mitchell Machine Learning Department Carnegie Mellon University October 11, 2012 Today: Computational Learning Theory Probably Approximately Coorrect (PAC) learning theorem

More information

VC-dimension of a context-dependent perceptron

VC-dimension of a context-dependent perceptron 1 VC-dimension of a context-dependent perceptron Piotr Ciskowski Institute of Engineering Cybernetics, Wroc law University of Technology, Wybrzeże Wyspiańskiego 27, 50 370 Wroc law, Poland cis@vectra.ita.pwr.wroc.pl

More information

Machine Learning

Machine Learning Machine Learning 10-601 Tom M. Mitchell Machine Learning Department Carnegie Mellon University October 11, 2012 Today: Computational Learning Theory Probably Approximately Coorrect (PAC) learning theorem

More information

Week 5: Logistic Regression & Neural Networks

Week 5: Logistic Regression & Neural Networks Week 5: Logistic Regression & Neural Networks Instructor: Sergey Levine 1 Summary: Logistic Regression In the previous lecture, we covered logistic regression. To recap, logistic regression models and

More information

ECE521 Lectures 9 Fully Connected Neural Networks

ECE521 Lectures 9 Fully Connected Neural Networks ECE521 Lectures 9 Fully Connected Neural Networks Outline Multi-class classification Learning multi-layer neural networks 2 Measuring distance in probability space We learnt that the squared L2 distance

More information

Artificial Neural Network Approach for Land Cover Classification of Fused Hyperspectral and Lidar Data

Artificial Neural Network Approach for Land Cover Classification of Fused Hyperspectral and Lidar Data Artificial Neural Network Approach for Land Cover Classification of Fused Hyperspectral and Lidar Data Paris Giampouras 1,2, Eleni Charou 1, and Anastasios Kesidis 3 1 Computational Intelligence Laboratory,

More information

Advanced statistical methods for data analysis Lecture 2

Advanced statistical methods for data analysis Lecture 2 Advanced statistical methods for data analysis Lecture 2 RHUL Physics www.pp.rhul.ac.uk/~cowan Universität Mainz Klausurtagung des GK Eichtheorien exp. Tests... Bullay/Mosel 15 17 September, 2008 1 Outline

More information

Linear Models for Regression

Linear Models for Regression Linear Models for Regression CSE 4309 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington 1 The Regression Problem Training data: A set of input-output

More information

Design Collocation Neural Network to Solve Singular Perturbed Problems with Initial Conditions

Design Collocation Neural Network to Solve Singular Perturbed Problems with Initial Conditions Article International Journal of Modern Engineering Sciences, 204, 3(): 29-38 International Journal of Modern Engineering Sciences Journal homepage:www.modernscientificpress.com/journals/ijmes.aspx ISSN:

More information

Neural Networks. Nicholas Ruozzi University of Texas at Dallas

Neural Networks. Nicholas Ruozzi University of Texas at Dallas Neural Networks Nicholas Ruozzi University of Texas at Dallas Handwritten Digit Recognition Given a collection of handwritten digits and their corresponding labels, we d like to be able to correctly classify

More information

Using Neural Networks for Identification and Control of Systems

Using Neural Networks for Identification and Control of Systems Using Neural Networks for Identification and Control of Systems Jhonatam Cordeiro Department of Industrial and Systems Engineering North Carolina A&T State University, Greensboro, NC 27411 jcrodrig@aggies.ncat.edu

More information

Ch.6 Deep Feedforward Networks (2/3)

Ch.6 Deep Feedforward Networks (2/3) Ch.6 Deep Feedforward Networks (2/3) 16. 10. 17. (Mon.) System Software Lab., Dept. of Mechanical & Information Eng. Woonggy Kim 1 Contents 6.3. Hidden Units 6.3.1. Rectified Linear Units and Their Generalizations

More information

A New Hybrid System for Recognition of Handwritten-Script

A New Hybrid System for Recognition of Handwritten-Script computing@tanet.edu.te.ua www.tanet.edu.te.ua/computing ISSN 177-69 A New Hybrid System for Recognition of Handwritten-Script Khalid Saeed 1) and Marek Tabdzki ) Faculty of Computer Science, Bialystok

More information

Simple neuron model Components of simple neuron

Simple neuron model Components of simple neuron Outline 1. Simple neuron model 2. Components of artificial neural networks 3. Common activation functions 4. MATLAB representation of neural network. Single neuron model Simple neuron model Components

More information

A Logarithmic Neural Network Architecture for Unbounded Non-Linear Function Approximation

A Logarithmic Neural Network Architecture for Unbounded Non-Linear Function Approximation 1 Introduction A Logarithmic Neural Network Architecture for Unbounded Non-Linear Function Approximation J Wesley Hines Nuclear Engineering Department The University of Tennessee Knoxville, Tennessee,

More information

Combination of M-Estimators and Neural Network Model to Analyze Inside/Outside Bark Tree Diameters

Combination of M-Estimators and Neural Network Model to Analyze Inside/Outside Bark Tree Diameters Combination of M-Estimators and Neural Network Model to Analyze Inside/Outside Bark Tree Diameters Kyriaki Kitikidou, Elias Milios, Lazaros Iliadis, and Minas Kaymakis Democritus University of Thrace,

More information

Machine Learning and Data Mining. Multi-layer Perceptrons & Neural Networks: Basics. Prof. Alexander Ihler

Machine Learning and Data Mining. Multi-layer Perceptrons & Neural Networks: Basics. Prof. Alexander Ihler + Machine Learning and Data Mining Multi-layer Perceptrons & Neural Networks: Basics Prof. Alexander Ihler Linear Classifiers (Perceptrons) Linear Classifiers a linear classifier is a mapping which partitions

More information

FEEDBACK GMDH-TYPE NEURAL NETWORK AND ITS APPLICATION TO MEDICAL IMAGE ANALYSIS OF LIVER CANCER. Tadashi Kondo and Junji Ueno

FEEDBACK GMDH-TYPE NEURAL NETWORK AND ITS APPLICATION TO MEDICAL IMAGE ANALYSIS OF LIVER CANCER. Tadashi Kondo and Junji Ueno International Journal of Innovative Computing, Information and Control ICIC International c 2012 ISSN 1349-4198 Volume 8, Number 3(B), March 2012 pp. 2285 2300 FEEDBACK GMDH-TYPE NEURAL NETWORK AND ITS

More information

Machine Learning Lecture 7

Machine Learning Lecture 7 Course Outline Machine Learning Lecture 7 Fundamentals (2 weeks) Bayes Decision Theory Probability Density Estimation Statistical Learning Theory 23.05.2016 Discriminative Approaches (5 weeks) Linear Discriminant

More information

Learning from Data: Regression

Learning from Data: Regression November 3, 2005 http://www.anc.ed.ac.uk/ amos/lfd/ Classification or Regression? Classification: want to learn a discrete target variable. Regression: want to learn a continuous target variable. Linear

More information

RAINFALL RUNOFF MODELING USING SUPPORT VECTOR REGRESSION AND ARTIFICIAL NEURAL NETWORKS

RAINFALL RUNOFF MODELING USING SUPPORT VECTOR REGRESSION AND ARTIFICIAL NEURAL NETWORKS CEST2011 Rhodes, Greece Ref no: XXX RAINFALL RUNOFF MODELING USING SUPPORT VECTOR REGRESSION AND ARTIFICIAL NEURAL NETWORKS D. BOTSIS1 1, P. LATINOPOULOS 2 and K. DIAMANTARAS 3 1&2 Department of Civil

More information

Support Vector Machines

Support Vector Machines Support Vector Machines INFO-4604, Applied Machine Learning University of Colorado Boulder September 28, 2017 Prof. Michael Paul Today Two important concepts: Margins Kernels Large Margin Classification

More information

Linear Models for Classification

Linear Models for Classification Linear Models for Classification Oliver Schulte - CMPT 726 Bishop PRML Ch. 4 Classification: Hand-written Digit Recognition CHINE INTELLIGENCE, VOL. 24, NO. 24, APRIL 2002 x i = t i = (0, 0, 0, 1, 0, 0,

More information

Acomplex-valued harmonic with a time-varying phase is a

Acomplex-valued harmonic with a time-varying phase is a IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 46, NO. 9, SEPTEMBER 1998 2315 Instantaneous Frequency Estimation Using the Wigner Distribution with Varying and Data-Driven Window Length Vladimir Katkovnik,

More information

Load Forecasting Using Artificial Neural Networks and Support Vector Regression

Load Forecasting Using Artificial Neural Networks and Support Vector Regression Proceedings of the 7th WSEAS International Conference on Power Systems, Beijing, China, September -7, 2007 3 Load Forecasting Using Artificial Neural Networks and Support Vector Regression SILVIO MICHEL

More information

ESTIMATING THE ACTIVATION FUNCTIONS OF AN MLP-NETWORK

ESTIMATING THE ACTIVATION FUNCTIONS OF AN MLP-NETWORK ESTIMATING THE ACTIVATION FUNCTIONS OF AN MLP-NETWORK P.V. Vehviläinen, H.A.T. Ihalainen Laboratory of Measurement and Information Technology Automation Department Tampere University of Technology, FIN-,

More information

Multilayer Perceptrons (MLPs)

Multilayer Perceptrons (MLPs) CSE 5526: Introduction to Neural Networks Multilayer Perceptrons (MLPs) 1 Motivation Multilayer networks are more powerful than singlelayer nets Example: XOR problem x 2 1 AND x o x 1 x 2 +1-1 o x x 1-1

More information

Content. Learning. Regression vs Classification. Regression a.k.a. function approximation and Classification a.k.a. pattern recognition

Content. Learning. Regression vs Classification. Regression a.k.a. function approximation and Classification a.k.a. pattern recognition Content Andrew Kusiak Intelligent Systems Laboratory 239 Seamans Center The University of Iowa Iowa City, IA 52242-527 andrew-kusiak@uiowa.edu http://www.icaen.uiowa.edu/~ankusiak Introduction to learning

More information

SUPPORT VECTOR REGRESSION WITH A GENERALIZED QUADRATIC LOSS

SUPPORT VECTOR REGRESSION WITH A GENERALIZED QUADRATIC LOSS SUPPORT VECTOR REGRESSION WITH A GENERALIZED QUADRATIC LOSS Filippo Portera and Alessandro Sperduti Dipartimento di Matematica Pura ed Applicata Universit a di Padova, Padova, Italy {portera,sperduti}@math.unipd.it

More information

On the complexity of shallow and deep neural network classifiers

On the complexity of shallow and deep neural network classifiers On the complexity of shallow and deep neural network classifiers Monica Bianchini and Franco Scarselli Department of Information Engineering and Mathematics University of Siena Via Roma 56, I-53100, Siena,

More information

Lecture 5: Logistic Regression. Neural Networks

Lecture 5: Logistic Regression. Neural Networks Lecture 5: Logistic Regression. Neural Networks Logistic regression Comparison with generative models Feed-forward neural networks Backpropagation Tricks for training neural networks COMP-652, Lecture

More information

Machine Learning. VC Dimension and Model Complexity. Eric Xing , Fall 2015

Machine Learning. VC Dimension and Model Complexity. Eric Xing , Fall 2015 Machine Learning 10-701, Fall 2015 VC Dimension and Model Complexity Eric Xing Lecture 16, November 3, 2015 Reading: Chap. 7 T.M book, and outline material Eric Xing @ CMU, 2006-2015 1 Last time: PAC and

More information

Artificial Neural Networks (ANN) Xiaogang Su, Ph.D. Department of Mathematical Science University of Texas at El Paso

Artificial Neural Networks (ANN) Xiaogang Su, Ph.D. Department of Mathematical Science University of Texas at El Paso Artificial Neural Networks (ANN) Xiaogang Su, Ph.D. Department of Mathematical Science University of Texas at El Paso xsu@utep.edu Fall, 2018 Outline Introduction A Brief History ANN Architecture Terminology

More information

Classifier Complexity and Support Vector Classifiers

Classifier Complexity and Support Vector Classifiers Classifier Complexity and Support Vector Classifiers Feature 2 6 4 2 0 2 4 6 8 RBF kernel 10 10 8 6 4 2 0 2 4 6 Feature 1 David M.J. Tax Pattern Recognition Laboratory Delft University of Technology D.M.J.Tax@tudelft.nl

More information

Neural Networks. Prof. Dr. Rudolf Kruse. Computational Intelligence Group Faculty for Computer Science

Neural Networks. Prof. Dr. Rudolf Kruse. Computational Intelligence Group Faculty for Computer Science Neural Networks Prof. Dr. Rudolf Kruse Computational Intelligence Group Faculty for Computer Science kruse@iws.cs.uni-magdeburg.de Rudolf Kruse Neural Networks 1 Supervised Learning / Support Vector Machines

More information

Artificial Neural Network and Fuzzy Logic

Artificial Neural Network and Fuzzy Logic Artificial Neural Network and Fuzzy Logic 1 Syllabus 2 Syllabus 3 Books 1. Artificial Neural Networks by B. Yagnanarayan, PHI - (Cover Topologies part of unit 1 and All part of Unit 2) 2. Neural Networks

More information

A New Weight Initialization using Statistically Resilient Method and Moore-Penrose Inverse Method for SFANN

A New Weight Initialization using Statistically Resilient Method and Moore-Penrose Inverse Method for SFANN A New Weight Initialization using Statistically Resilient Method and Moore-Penrose Inverse Method for SFANN Apeksha Mittal, Amit Prakash Singh and Pravin Chandra University School of Information and Communication

More information

4. Multilayer Perceptrons

4. Multilayer Perceptrons 4. Multilayer Perceptrons This is a supervised error-correction learning algorithm. 1 4.1 Introduction A multilayer feedforward network consists of an input layer, one or more hidden layers, and an output

More information

Application of Artificial Neural Networks in Evaluation and Identification of Electrical Loss in Transformers According to the Energy Consumption

Application of Artificial Neural Networks in Evaluation and Identification of Electrical Loss in Transformers According to the Energy Consumption Application of Artificial Neural Networks in Evaluation and Identification of Electrical Loss in Transformers According to the Energy Consumption ANDRÉ NUNES DE SOUZA, JOSÉ ALFREDO C. ULSON, IVAN NUNES

More information

SVAN 2016 Mini Course: Stochastic Convex Optimization Methods in Machine Learning

SVAN 2016 Mini Course: Stochastic Convex Optimization Methods in Machine Learning SVAN 2016 Mini Course: Stochastic Convex Optimization Methods in Machine Learning Mark Schmidt University of British Columbia, May 2016 www.cs.ubc.ca/~schmidtm/svan16 Some images from this lecture are

More information

Oil Field Production using Machine Learning. CS 229 Project Report

Oil Field Production using Machine Learning. CS 229 Project Report Oil Field Production using Machine Learning CS 229 Project Report Sumeet Trehan, Energy Resources Engineering, Stanford University 1 Introduction Effective management of reservoirs motivates oil and gas

More information

CSE 417T: Introduction to Machine Learning. Lecture 11: Review. Henry Chai 10/02/18

CSE 417T: Introduction to Machine Learning. Lecture 11: Review. Henry Chai 10/02/18 CSE 417T: Introduction to Machine Learning Lecture 11: Review Henry Chai 10/02/18 Unknown Target Function!: # % Training data Formal Setup & = ( ), + ),, ( -, + - Learning Algorithm 2 Hypothesis Set H

More information

Feed-forward Network Functions

Feed-forward Network Functions Feed-forward Network Functions Sargur Srihari Topics 1. Extension of linear models 2. Feed-forward Network Functions 3. Weight-space symmetries 2 Recap of Linear Models Linear Models for Regression, Classification

More information

Linear Models for Regression CS534

Linear Models for Regression CS534 Linear Models for Regression CS534 Prediction Problems Predict housing price based on House size, lot size, Location, # of rooms Predict stock price based on Price history of the past month Predict the

More information

Statistical Learning Reading Assignments

Statistical Learning Reading Assignments Statistical Learning Reading Assignments S. Gong et al. Dynamic Vision: From Images to Face Recognition, Imperial College Press, 2001 (Chapt. 3, hard copy). T. Evgeniou, M. Pontil, and T. Poggio, "Statistical

More information

Are Rosenblatt multilayer perceptrons more powerfull than sigmoidal multilayer perceptrons? From a counter example to a general result

Are Rosenblatt multilayer perceptrons more powerfull than sigmoidal multilayer perceptrons? From a counter example to a general result Are Rosenblatt multilayer perceptrons more powerfull than sigmoidal multilayer perceptrons? From a counter example to a general result J. Barahona da Fonseca Department of Electrical Engineering, Faculty

More information

A STATE-SPACE NEURAL NETWORK FOR MODELING DYNAMICAL NONLINEAR SYSTEMS

A STATE-SPACE NEURAL NETWORK FOR MODELING DYNAMICAL NONLINEAR SYSTEMS A STATE-SPACE NEURAL NETWORK FOR MODELING DYNAMICAL NONLINEAR SYSTEMS Karima Amoura Patrice Wira and Said Djennoune Laboratoire CCSP Université Mouloud Mammeri Tizi Ouzou Algeria Laboratoire MIPS Université

More information

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Cheng Soon Ong & Christian Walder. Canberra February June 2018 Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 254 Part V

More information

ARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92

ARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92 ARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92 BIOLOGICAL INSPIRATIONS Some numbers The human brain contains about 10 billion nerve cells (neurons) Each neuron is connected to the others through 10000

More information

Optimization Methods for Machine Learning (OMML)

Optimization Methods for Machine Learning (OMML) Optimization Methods for Machine Learning (OMML) 2nd lecture (2 slots) Prof. L. Palagi 16/10/2014 1 What is (not) Data Mining? By Namwar Rizvi - Ad Hoc Query: ad Hoc queries just examines the current data

More information

A Novel Activity Detection Method

A Novel Activity Detection Method A Novel Activity Detection Method Gismy George P.G. Student, Department of ECE, Ilahia College of,muvattupuzha, Kerala, India ABSTRACT: This paper presents an approach for activity state recognition of

More information

Intelligent Modular Neural Network for Dynamic System Parameter Estimation

Intelligent Modular Neural Network for Dynamic System Parameter Estimation Intelligent Modular Neural Network for Dynamic System Parameter Estimation Andrzej Materka Technical University of Lodz, Institute of Electronics Stefanowskiego 18, 9-537 Lodz, Poland Abstract: A technique

More information

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18 CSE 417T: Introduction to Machine Learning Final Review Henry Chai 12/4/18 Overfitting Overfitting is fitting the training data more than is warranted Fitting noise rather than signal 2 Estimating! "#$

More information

Discriminative Models

Discriminative Models No.5 Discriminative Models Hui Jiang Department of Electrical Engineering and Computer Science Lassonde School of Engineering York University, Toronto, Canada Outline Generative vs. Discriminative models

More information

Forecasting Crude Oil Price Using Neural Networks

Forecasting Crude Oil Price Using Neural Networks CMU. Journal (2006) Vol. 5(3) 377 Forecasting Crude Oil Price Using Neural Networks Komsan Suriya * Faculty of Economics, Chiang Mai University, Chiang Mai 50200, Thailand *Corresponding author. E-mail:

More information

VC dimension, Model Selection and Performance Assessment for SVM and Other Machine Learning Algorithms

VC dimension, Model Selection and Performance Assessment for SVM and Other Machine Learning Algorithms 03/Feb/2010 VC dimension, Model Selection and Performance Assessment for SVM and Other Machine Learning Algorithms Presented by Andriy Temko Department of Electrical and Electronic Engineering Page 2 of

More information

Discussion About Nonlinear Time Series Prediction Using Least Squares Support Vector Machine

Discussion About Nonlinear Time Series Prediction Using Least Squares Support Vector Machine Commun. Theor. Phys. (Beijing, China) 43 (2005) pp. 1056 1060 c International Academic Publishers Vol. 43, No. 6, June 15, 2005 Discussion About Nonlinear Time Series Prediction Using Least Squares Support

More information

y(n) Time Series Data

y(n) Time Series Data Recurrent SOM with Local Linear Models in Time Series Prediction Timo Koskela, Markus Varsta, Jukka Heikkonen, and Kimmo Kaski Helsinki University of Technology Laboratory of Computational Engineering

More information

Multilayer Perceptron

Multilayer Perceptron Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Single Perceptron 3 Boolean Function Learning 4

More information

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Cheng Soon Ong & Christian Walder. Canberra February June 2018 Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 Outlines Overview Introduction Linear Algebra Probability Linear Regression

More information

Logistic Regression Review Fall 2012 Recitation. September 25, 2012 TA: Selen Uguroglu

Logistic Regression Review Fall 2012 Recitation. September 25, 2012 TA: Selen Uguroglu Logistic Regression Review 10-601 Fall 2012 Recitation September 25, 2012 TA: Selen Uguroglu!1 Outline Decision Theory Logistic regression Goal Loss function Inference Gradient Descent!2 Training Data

More information

CSC 578 Neural Networks and Deep Learning

CSC 578 Neural Networks and Deep Learning CSC 578 Neural Networks and Deep Learning Fall 2018/19 3. Improving Neural Networks (Some figures adapted from NNDL book) 1 Various Approaches to Improve Neural Networks 1. Cost functions Quadratic Cross

More information

ESTIMATION OF HOURLY MEAN AMBIENT TEMPERATURES WITH ARTIFICIAL NEURAL NETWORKS 1. INTRODUCTION

ESTIMATION OF HOURLY MEAN AMBIENT TEMPERATURES WITH ARTIFICIAL NEURAL NETWORKS 1. INTRODUCTION Mathematical and Computational Applications, Vol. 11, No. 3, pp. 215-224, 2006. Association for Scientific Research ESTIMATION OF HOURLY MEAN AMBIENT TEMPERATURES WITH ARTIFICIAL NEURAL NETWORKS Ömer Altan

More information

Vapnik-Chervonenkis Dimension of Neural Nets

Vapnik-Chervonenkis Dimension of Neural Nets Vapnik-Chervonenkis Dimension of Neural Nets Peter L. Bartlett BIOwulf Technologies and University of California at Berkeley Department of Statistics 367 Evans Hall, CA 94720-3860, USA bartlett@stat.berkeley.edu

More information

When can Deep Networks avoid the curse of dimensionality and other theoretical puzzles

When can Deep Networks avoid the curse of dimensionality and other theoretical puzzles When can Deep Networks avoid the curse of dimensionality and other theoretical puzzles Tomaso Poggio, MIT, CBMM Astar CBMM s focus is the Science and the Engineering of Intelligence We aim to make progress

More information

Neural Network Control in a Wastewater Treatment Plant

Neural Network Control in a Wastewater Treatment Plant Neural Network Control in a Wastewater Treatment Plant Miguel A. Jaramillo 1 ; Juan C. Peguero 2, Enrique Martínez de Salazar 1, Montserrat García del alle 1 ( 1 )Escuela de Ingenierías Industriales. (

More information

Midterm Review CS 7301: Advanced Machine Learning. Vibhav Gogate The University of Texas at Dallas

Midterm Review CS 7301: Advanced Machine Learning. Vibhav Gogate The University of Texas at Dallas Midterm Review CS 7301: Advanced Machine Learning Vibhav Gogate The University of Texas at Dallas Supervised Learning Issues in supervised learning What makes learning hard Point Estimation: MLE vs Bayesian

More information

A Training-time Analysis of Robustness in Feed-Forward Neural Networks

A Training-time Analysis of Robustness in Feed-Forward Neural Networks A Training-time Analysis of Robustness in Feed-Forward Neural Networks Cesare Alippi Dipartimento di Elettronica e Informazione Politecnico di Milano Milano, Italy E-mail: alippi@elet.polimi Daniele Sana,

More information

Novel determination of dierential-equation solutions: universal approximation method

Novel determination of dierential-equation solutions: universal approximation method Journal of Computational and Applied Mathematics 146 (2002) 443 457 www.elsevier.com/locate/cam Novel determination of dierential-equation solutions: universal approximation method Thananchai Leephakpreeda

More information

Neural Networks Lecturer: J. Matas Authors: J. Matas, B. Flach, O. Drbohlav

Neural Networks Lecturer: J. Matas Authors: J. Matas, B. Flach, O. Drbohlav Neural Networks 30.11.2015 Lecturer: J. Matas Authors: J. Matas, B. Flach, O. Drbohlav 1 Talk Outline Perceptron Combining neurons to a network Neural network, processing input to an output Learning Cost

More information

Linear Dependency Between and the Input Noise in -Support Vector Regression

Linear Dependency Between and the Input Noise in -Support Vector Regression 544 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 14, NO. 3, MAY 2003 Linear Dependency Between the Input Noise in -Support Vector Regression James T. Kwok Ivor W. Tsang Abstract In using the -support vector

More information

Resampling techniques for statistical modeling

Resampling techniques for statistical modeling Resampling techniques for statistical modeling Gianluca Bontempi Département d Informatique Boulevard de Triomphe - CP 212 http://www.ulb.ac.be/di Resampling techniques p.1/33 Beyond the empirical error

More information

Neural network modelling of reinforced concrete beam shear capacity

Neural network modelling of reinforced concrete beam shear capacity icccbe 2010 Nottingham University Press Proceedings of the International Conference on Computing in Civil and Building Engineering W Tizani (Editor) Neural network modelling of reinforced concrete beam

More information

10-701/ Machine Learning, Fall

10-701/ Machine Learning, Fall 0-70/5-78 Machine Learning, Fall 2003 Homework 2 Solution If you have questions, please contact Jiayong Zhang .. (Error Function) The sum-of-squares error is the most common training

More information

EM-algorithm for Training of State-space Models with Application to Time Series Prediction

EM-algorithm for Training of State-space Models with Application to Time Series Prediction EM-algorithm for Training of State-space Models with Application to Time Series Prediction Elia Liitiäinen, Nima Reyhani and Amaury Lendasse Helsinki University of Technology - Neural Networks Research

More information

Machine Learning. 7. Logistic and Linear Regression

Machine Learning. 7. Logistic and Linear Regression Sapienza University of Rome, Italy - Machine Learning (27/28) University of Rome La Sapienza Master in Artificial Intelligence and Robotics Machine Learning 7. Logistic and Linear Regression Luca Iocchi,

More information

Vapnik-Chervonenkis Dimension of Neural Nets

Vapnik-Chervonenkis Dimension of Neural Nets P. L. Bartlett and W. Maass: Vapnik-Chervonenkis Dimension of Neural Nets 1 Vapnik-Chervonenkis Dimension of Neural Nets Peter L. Bartlett BIOwulf Technologies and University of California at Berkeley

More information

Statistical learning theory, Support vector machines, and Bioinformatics

Statistical learning theory, Support vector machines, and Bioinformatics 1 Statistical learning theory, Support vector machines, and Bioinformatics Jean-Philippe.Vert@mines.org Ecole des Mines de Paris Computational Biology group ENS Paris, november 25, 2003. 2 Overview 1.

More information

MODULAR ECHO STATE NEURAL NETWORKS IN TIME SERIES PREDICTION

MODULAR ECHO STATE NEURAL NETWORKS IN TIME SERIES PREDICTION Computing and Informatics, Vol. 30, 2011, 321 334 MODULAR ECHO STATE NEURAL NETWORKS IN TIME SERIES PREDICTION Štefan Babinec, Jiří Pospíchal Department of Mathematics Faculty of Chemical and Food Technology

More information

Deep Feedforward Networks

Deep Feedforward Networks Deep Feedforward Networks Liu Yang March 30, 2017 Liu Yang Short title March 30, 2017 1 / 24 Overview 1 Background A general introduction Example 2 Gradient based learning Cost functions Output Units 3

More information

Support Vector Machines vs Multi-Layer. Perceptron in Particle Identication. DIFI, Universita di Genova (I) INFN Sezione di Genova (I) Cambridge (US)

Support Vector Machines vs Multi-Layer. Perceptron in Particle Identication. DIFI, Universita di Genova (I) INFN Sezione di Genova (I) Cambridge (US) Support Vector Machines vs Multi-Layer Perceptron in Particle Identication N.Barabino 1, M.Pallavicini 2, A.Petrolini 1;2, M.Pontil 3;1, A.Verri 4;3 1 DIFI, Universita di Genova (I) 2 INFN Sezione di Genova

More information

HOPFIELD neural networks (HNNs) are a class of nonlinear

HOPFIELD neural networks (HNNs) are a class of nonlinear IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II: EXPRESS BRIEFS, VOL. 52, NO. 4, APRIL 2005 213 Stochastic Noise Process Enhancement of Hopfield Neural Networks Vladimir Pavlović, Member, IEEE, Dan Schonfeld,

More information

Midterm Review CS 6375: Machine Learning. Vibhav Gogate The University of Texas at Dallas

Midterm Review CS 6375: Machine Learning. Vibhav Gogate The University of Texas at Dallas Midterm Review CS 6375: Machine Learning Vibhav Gogate The University of Texas at Dallas Machine Learning Supervised Learning Unsupervised Learning Reinforcement Learning Parametric Y Continuous Non-parametric

More information

Bayesian Machine Learning

Bayesian Machine Learning Bayesian Machine Learning Andrew Gordon Wilson ORIE 6741 Lecture 2: Bayesian Basics https://people.orie.cornell.edu/andrew/orie6741 Cornell University August 25, 2016 1 / 17 Canonical Machine Learning

More information

Lecture 10. Neural networks and optimization. Machine Learning and Data Mining November Nando de Freitas UBC. Nonlinear Supervised Learning

Lecture 10. Neural networks and optimization. Machine Learning and Data Mining November Nando de Freitas UBC. Nonlinear Supervised Learning Lecture 0 Neural networks and optimization Machine Learning and Data Mining November 2009 UBC Gradient Searching for a good solution can be interpreted as looking for a minimum of some error (loss) function

More information

NONLINEAR BLACK BOX MODELING OF A LEAD ACID BATTERY USING HAMMERSTEIN-WIENER MODEL

NONLINEAR BLACK BOX MODELING OF A LEAD ACID BATTERY USING HAMMERSTEIN-WIENER MODEL NONLINEAR BLACK BOX MODELING OF A LEAD ACID BATTERY USING HAMMERSTEIN-WIENER MODEL 1 EL MEHDI LAADISSI, 2 EL FILALI ANAS, 3 MALIKA ZAZI 1, 2, 3 Laboratory LM2PI, ENSET, Mohamed V University Rabat, Morocco

More information

Discriminative Models

Discriminative Models No.5 Discriminative Models Hui Jiang Department of Electrical Engineering and Computer Science Lassonde School of Engineering York University, Toronto, Canada Outline Generative vs. Discriminative models

More information

1162 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 11, NO. 5, SEPTEMBER The Evidence Framework Applied to Support Vector Machines

1162 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 11, NO. 5, SEPTEMBER The Evidence Framework Applied to Support Vector Machines 1162 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 11, NO. 5, SEPTEMBER 2000 Brief Papers The Evidence Framework Applied to Support Vector Machines James Tin-Yau Kwok Abstract In this paper, we show that

More information

Ways to make neural networks generalize better

Ways to make neural networks generalize better Ways to make neural networks generalize better Seminar in Deep Learning University of Tartu 04 / 10 / 2014 Pihel Saatmann Topics Overview of ways to improve generalization Limiting the size of the weights

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning Neural Networks Varun Chandola x x 5 Input Outline Contents February 2, 207 Extending Perceptrons 2 Multi Layered Perceptrons 2 2. Generalizing to Multiple Labels.................

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning Vapnik Chervonenkis Theory Barnabás Póczos Empirical Risk and True Risk 2 Empirical Risk Shorthand: True risk of f (deterministic): Bayes risk: Let us use the empirical

More information

Reminders. Thought questions should be submitted on eclass. Please list the section related to the thought question

Reminders. Thought questions should be submitted on eclass. Please list the section related to the thought question Linear regression Reminders Thought questions should be submitted on eclass Please list the section related to the thought question If it is a more general, open-ended question not exactly related to a

More information