Defining Feedforward Network Architecture. net = newff([pn],[s1 S2... SN],{TF1 TF2... TFN},BTF,LF,PF);
|
|
- Berniece Justina Cobb
- 5 years ago
- Views:
Transcription
1 Appendix D MATLAB Programs for Neural Systems D.. Defining Feedforward Network Architecture Feedforward networks often have one or more hidden layers of sigmoid neurons followed by an output layer of linear neurons. Multiple layers of neurons with nonlinear transfer functions allow the network to learn nonlinear and linear relationships between input and output vectors. The function newff() creates a feedforward backpropagation network architecture with desired number of layers and neurons. The general form of use of the function is given below, which returns an N-layer feedforward backpropagation network object: net = newff([pn],[s S2... SN],{TF TF2... TFN},BTF,LF,PF); where the first input PN is an N 2 matrix of minimum and maximum values for N input elements. S S2... SN are the sizes (number of neurons) of the layers of the network architecture. TFi is the transfer function of the ith layer; the default is tansig. The transfer functions TFi can be any differentiable transfer function such as tansig, logsig or purelin. BTF is the backpropagation network training function; the default is trainlm. Different training functions with their features are described in Section of Chapter 4. LF is the backpropagation weight/bias learning function with gradient descent, such as learngd, learngdm. The default is learngdm. The function learngdm is used to calculate the weight change dw for a given neuron from the neuron s input P and error E. Learning occurs according to learngdm s learning parameters such as the weight (or bias) W, learning rate and momentum constant, according to gradient descent with momentum and returns weight change and new learning states. PF is the performance function such as mse (mean squared error), mae (mean absolute error) and msereg (mean squared error with regularization). The default is mse. For example: net=newff([- 2; 5],[3,],{'tansig','purelin'},'traingd', 'learngdm', 'mae'); The function creates a two-input single-output feedforward network with single hidden layer. The first input [ 2; 5] specifies the minimum and maximum values for each of the input vectors. The second input is an array containing the sizes of each layer, i.e., the network Computational Intelligence: Synergies of Fuzzy Logic, Neural Networks and Evolutionary Computing, First Edition. Nazmul Siddique and Hojjat Adeli. 23 John Wiley & Sons, Ltd. Published 23 by John Wiley & Sons, Ltd.
2 462 Computational Intelligence has 3 neurons in the hidden layer and neuron in the output layer. The third input is a cell array containing the names of the transfer functions to be used in each layer, i.e., tansig for hidden layer and purelin (linear) activation function for output layer. There are other different activation functions with distinct features, such as logsig, hardlim. The final input contains the name of the training function to be used. traingd is one of the training functions used by the network. newff() will also automatically initialize the weights and biases of the network. D... Creating RBF Network Architecture In an RBF network, there can be a maximum of M inputs and a maximum of N radial basis neurons in the hidden layer. There are no weights between inputs and hidden neurons. Each radial basis neuron is connected to the output neuron through the weight matrix W, which has to be learned. Using the MATLAB R functions, the architecture of an RBF network with m =, 2, 3,...,M input elements and n =, 2, 3,...,N radial basis neurons (in the hidden layer) can be created. All details of designing a radial basis function network are built into the design functions newrbe() and newrb(), and their outputs can be obtained with sim(). The functions are called in the following way: net = newrbe(p, T, Spread); The function newrbe() takes matrices of input vectors P and target vectors T, and a spread for the radial basis layer, and returns a network with weights and biases such that the outputs are exactly T when the inputs are P. The value for the spread constant should be larger than the distance between adjacent input vectors, so as to get a good generalization, but smaller than the distance across the whole input space. The function newrbe() creates as many radial basis neurons as there are input vectors in P. The drawback to newrbe() is that it produces a network with as many hidden neurons as there are input vectors. For this reason, newrbe() does not return an acceptable solution when many input vectors are needed to properly define a network, as is typically the case (Demuth and Beale, 2). newrb() is a more efficient design function, which creates a radial basis network one neuron at a time. Neurons are added to the network until the sum-squared error falls beneath an error goal or a maximum number of neurons have been reached. The call for this function is net = newrb(p, T, Goal, Spread); The function newrb() takes matrices of input vectors P and target vectors T, and design parameters goal and spread for radial basis layer, and returns the desired network with weights and biases such that the outputs are exactly T when the inputs are P. The design method of newrb() is similar to that of newrbe(). The difference is that newrb() creates neurons one at a time. The error of the new network is checked, and if low enough newrb() is finished. Otherwise the next neuron is added. This procedure is repeated until the error goal is met or the maximum number of neurons is reached. Thus, newrbe() creates a network with zero error on training vectors. The only condition required is to make sure that SPREAD is large enough so that the active input regions of the radbas neurons overlap enough so that several radbas neurons always have fairly large outputs at any given moment. This makes the network function smoother and results in better generalization for new input vectors occurring between input vectors used in the design. (However, SPREAD should not be so large that each neuron is effectively responding in the same large area of the input space.)
3 Appendix D: MATLAB Programs for Neural Systems 463 RBF networks, even when designed effectively with newrbe(), tend to have many times more neurons than a comparable MLP network with tansig or logsig neurons in the hidden layer. This is because sigmoid neurons can have outputs over a large region of the input space, while RBF neurons only respond to relatively small regions of the input space. The result is that the larger the input space (in terms of number of inputs, and the ranges those inputs vary over) the more RBF neurons are required. On the other hand, designing an RBF network often takes much less time than training a sigmoid/linear network, and can sometimes result in fewer neurons being used. D...2 Creating GRNN Network Architecture A generalized regression neural network is often used for function approximation. A GRNN network with m =, 2, 3,..., M input elements and n =, 2, 3,..., N radial basis neurons (in the hidden layer) can be created using the function newgrnn() where the first layer is just like that of newrbe() or newrb() but has a slightly different second layer. It has as many neurons as there are input/target vectors in P. Specifically, the first-layer weights are set to P. The bias b is set to a column vector of.8326/spread. The user chooses Spread, the distance an input vector must be from a neuron s weight vector to be.5. Each neuron s weighted input is the distance between the input vector and its weight vector. The second layer also has as many neurons as input/target vectors, but there the weights are set to target T. The function is called in the following way: net = newgrnn(p, T, Spread); The function newgrnn() takes matrices of input vectors P and target vectors T, and a spread for the radial basis layer, and returns a network with weights and biases such that the outputs are exactly T when the inputs are P. The value for spread constant should be larger than the distance between adjacent input vectors, so as to get good generalization, but smaller than the distance across the whole input space. To fit data closely, a smaller spread is suggested, i.e., smaller than the typical distance between input vectors. To fit the data more smoothly, a larger spread is to be chosen. A larger spread leads to a large area around the input vector where layer neurons will respond with significant outputs. Therefore, if the spread is small the radial basis function is very steep, so that the neuron with weight vector closest to the input will have a much larger output than other neurons. The network tends to respond with the target vector associated with the nearest design input vector. As the spread becomes larger the radial basis function s slope becomes smoother and several neurons can respond to an input vector. The network then acts as if it is taking a weighted average between target vectors whose design input vectors are closest to the new input vector. As the spread becomes larger, more and more neurons contribute to the average, with the result that the network function becomes smoother. D...3 Creating PRNN Network Architecture A PNN network can be created by calling the function in the following way: net = newpnn(p, T, Spread); The function newpnn() takes matrices of input vectors P and target vectors T, and a spread for the radial basis layer, and returns a network with weights and biases such that the outputs
4 464 Computational Intelligence are exactly T when the inputs are P. If the spread is near zero, the network will act as a nearest-neighbour classifier. As the spread becomes larger, the designed network will take into account several nearby design vectors. Although the PNN was derived from the same mathematical merits and similarities to those of RBF and GRNN networks, after defining the architecture it is found to be more appropriate for classification problems rather than prediction or approximation problems. Probabilistic neural networks can be used for classification problems. When an input is presented, the first layer computes distances from the input vector to the training input vectors and produces a vector whose elements indicate how close the input is to a training input. The second layer sums these contributions for each class of inputs to produce as its net output a vector of probabilities. Finally, a complete transfer function on the output of the second layer picks the maximum of these probabilities, and produces a for that class and a for the other classes. D..2 Training Networks Different backpropagation training algorithms are available as functions in MATLAB R. They have their own features and advantages. Some of the most widely used functions are discussed briefly. traingd basic gradient descent learning algorithm. It has slow response but can be used in incremental mode training. traingdm gradient descent with momentum. It is generally faster than traingd and can be used in incremental mode training. traingdx adaptive learning rate. It has faster training time than traingd but can only be used in batch mode training. trainrp resilient backpropagation. It is a simple batch mode training algorithm with fast convergence and minimal storage requirements. trainlm Levenberg Marquart algorithm. It is a faster training algorithm for networks of moderate size. It has a memory reduction feature for use when the training set is large. There are several parameters associated with training algorithms. The parameters are learning rate, error goal, epochs and show. These parameters are defined as: net.trainparam.lr - specifies learning rate net.trainparam.goal - specifies error goal net.trainparam.epochs - specifies the number of iterations net.trainparam.show - displays status for every show. Once the network has been defined and the parameters are set, the network can be trained using the function train() as [net, tr] = train(net, P, T) where net is the network object, tr contains information about the progress of training, P and T are the input and output vectors, respectively. Typically, one epoch of training is defined as a
5 Appendix D: MATLAB Programs for Neural Systems 465 single presentation of all input vectors to the network. The network is then updated according to the results of all those presentations. Training occurs until a maximum number of epochs occur, the performance goal is met or any other stopping condition of the function is met. For example: net.trainparam.lr =.5; net.trainparam.goal =.; net.trainparam.epochs =; net.trainparam.show = 25; [net,tr]=train(net, P, T) The network will be trained using the input and output data P and T, respectively for up to epochs or when an error goal of. is reached. D..3 Simulating Networks The function sim() simulates a network. It takes the network input P and the network objects net and returns the network output ŷ. A single matrix of concurrent vector is presented to the network and the network produces a single matrix of concurrent vector as output: ŷ= sim(net, P) D..4 Creating Neural Network Subsystem Once the network has been trained and tested with training and checking data, a Simulink R model can be created using the MATLAB R function gensim(). The function gensim() generates block descriptions of networks so that it can simulate the neural network in Simulink R.The function is called in the following way: gensim(net, st) gensim() takes these inputs net ˆ= neural network defined either in an M-file or NN Toolbox and st ˆ= sample time and creates a Simulink R system containing a block which simulates a neural network with a specified sampling time: gensim(net,st) The second argument to gensim() determines the sample time, which is normally chosen to be some positive real value. If the network has no delays associated with its input weights or layer weights, this value is set to. For example: gensim(net,-) The value of the parameter st is, i.e., tells gensim to generate a network with continuous sampling. Example D..: Define a feedforward network, train and simulate with input data In this example, a two-layer feedforward network is created. The network s input ranges from
6 466 Computational Intelligence to. The first layer has five neurons with tansig function; the second layer has one neuron with linear function. The training function traingd is used to train the network. %Chapter 4 Example D.. %Create a feedforward NN and train with data set [P T] %Training set P = [ ]; T = [ ]; net = newff([ ],[5 ],{'tansig' 'purelin'},'traingd'); %Set network parameters as follows net.trainparam.lr =. %Learning rate net.trainparam.goal =. %Performance goal net.trainparam.epochs = 5 %This sets maximum number of epochs in a training net.trainparam.show = 25 %This displays training status after each 25 epoch net.trainparam.time = inf %Maximum time to train in seconds net.trainparam.min_grad=e- %Minimum performance gradient %Here the network is simulated and its output plotted against the targets. net = train(net,p,t); Y = sim(net,p); plot(p,t, '-k', P,Y, 'ok') Example D..2: Updating weights of a two-layer neural net A two-layer neural network with two inputs x = [x, x 2 ] and one output y is given by y = W2 T [ f ( [ ] [ ] W T x + b ) ] + b2, where W T =, b =, W T = [ ] and b 2 = [ 2.8]. Update the weights and biases of the network, simulate the network and plot the output surface over the grid [ 2, 2] [ 2, 2]: %Chapter 4 Example D..2 %A two layer NN is given by y=w2'[f(w'x+b)]+b2 %with W=[ ; ]; b=[-2.29; 3.67]; W2=[ ]; b2=[-2.8]; %Update NN with W, W2, b and b2 %Plot the NN output surface y as a function of x over grid [-2,2] x[-2,2] %Weights and bias W=[ ; ]; b=[-2.29; 3.67];
7 Appendix D: MATLAB Programs for Neural Systems 467 W2=[ ]; b2=[-2.8]; %Output surface [x, x2]=meshgrid(-2:.:2); %Compute NN output p=x(:); p2=x2(:); p=[p';p2']; %NN weights and bias %nnt2ff () updates NN with specified weights and biases net=nnt2ff(minmax(p),{w,w2},{b,b2},{'tansig', 'purelin'}); %Simulate a=sim(net,p); %Arrange results for mesh plot a=eye(4); a(:)=a'; mesh(x,x2,a); AZ=6, EL=3; view(az,el); xlabel('x'); ylabel('x2'); title('nn output surface for tansigmoid function') See Figure D.. for the result. Example D..3: Approximation of output surface In this example, an output surface of a nonlinear function is approximated. The nonlinear function is defined by f (x, y) = NN output surface for tansigmoid function 5 5 x 2 x2 2 Figure D.. Output surface of a two-layer network
8 468 Computational Intelligence sin (π x) cos (πy) with x [ 2, 2] and y [ 2, 2]. An MLP with 2 tansigmoidal neurons and linear neuron can approximate the function after training the network for 5 epochs: %Chapter 4 - Example D..3 %NN Function approximation [x, y]=meshgrid(-2:.:2); %Nonlinear function defined by z=sin(pi*x).*cos(pi*y); %Generate Input & Target data for i=:2 p(:,i)=4*(rand(2,)-.5); T(:,i)=sin(pi*p(2*i-))*cos(pi*p(2*i)); end % %Two-layer NN created with 2 tansig %and one purelin neuron net=newff(minmax(p), [2,], {'tansig', 'purelin'}, 'trainlm'); net.trainparam.show=5; net.trainparam.epochs=5; net.trainparam.goal=e-6; [net,tr]=train(net,p,t) %Simulate the net a=zeros(4,4); [x, y]=meshgrid(-2:.:2); for i=:68 a(i)=sim(net,[x(i);y(i)]); end figure() %Original nonlinear function subplot(,2,) mesh(x,y,z); title('original function graphics'); xlabel('<--x-->') ylabel('<--y-->') zlabel('<--z-->') AZ=5; EL=59; view(az,el) subplot(,2,2) mesh(x,y,a); title('nn approximated graphics') xlabel('<--x-->') ylabel('<--y-->') zlabel('<--z-->') AZ=5; EL=59; view(az,el)
9 Appendix D: MATLAB Programs for Neural Systems 469 Original function graphics NN approximated graphics <--z--> <--z--> 2 <--y--> 2 2 <--x--> <--x--> 2 <--y--> 2 (a) (b) Figure D..2 function Approximation of a nonlinear function. (a) Nonlinear function; (b) NN approximated See Figure D..2. Example D..4: Approximation of a nonlinear function In this example, a nonlinear function defined by the input/output data is approximated using a two-layer feedforward network. The network s weights and biases are shown after training. The plot will also show how good the approximation is: %Chapter 4 Example D..4 %Function approximation clear all; close all; %Training data:examplar input pattern and target output vector x=-:.:; y=[-.96,-.577,-.73,.377,.64,.66,.46,.34, , -.434, -.5, -.393, -.65,.99,.37,.396, ,.82, -.3, -.29, -.32]; %Define a NN and initialise weights net=newff(minmax(x), [7 ], {'tansig', 'purelin'},'trainlm'); %Output of NN with initial weights ycapl=sim(net,x); %Train the NN net.trainparam.epochs = 5 net.trainparam.goal =. net.trainparam.lr =. %net.trainparam.min_grad=e- %Maximum number of epochs to train %Performance goal %Learning rate %Minimum performance gradient
10 47 Computational Intelligence net.trainparam.show = net.trainparam.time = inf [net,tr]=train(net,x,y); %Epochs between displays %Maximum time to train in seconds %Output of NN figure() %Generalisation: input vector is different %from the one used for training x2=-:.:; ycap2=sim(net,x2); plot(x,ycap,x2,ycap2,'-',x,y,'o'); plot(x,ycapl,'--k', x2,ycap2,'-.k',x,y,'ok'); title('function approximation'); xlabel('x-values'); ylabel('y-values'); legend('before Training', 'After Training','Function'); %Show weights and biases of NN w=net.iw{,} bw=net.b{} v=net.lw{2,} bv=net.b{2} See Figure D Function approximation Before Training After Training Function.2 y-values x-values Figure D..3 Function approximation
11 Appendix D: MATLAB Programs for Neural Systems 47 T double Target P Input double double p{} y{} Neural Network double (2) y{} Figure D..4 NN block description for Simulink R Example D..5: Creating an NN subsystem for simulation In this example, a neural network block description is created for simulation to be used under Simulink R.Todothis, an NN is trained with a set of input/output data and simulated. Once this is done, an NN block description is created using the gensim() function: %Chapter 4 Example D..5 %Training data set p=[-:.5:]; %noisy sine wave t=sin(2*pi*p)+.*randn(size(p)); net=newff([- ],[2,],{'tansig','purelin'},'traingdx'); net.trainparam.show=5; net.trainparam.epochs=3; pt=p*.979; y=sim(net,pt) plot(p,t,'-',p,y,'o') %generate NN block description gensim(net,-) The generated NN Simulink R block description is shown in Figure D..4. The Neural Network Toolbox provides three popular neural network Simulink R blocks for prediction and control that have been applied to many applications: Model Predictive Control, NARMA-L2 (or Feedback Linearization) Control and Model Reference Control. These Simulink R control blocks are discussed in detail in Chapter 5.
Instruction Sheet for SOFT COMPUTING LABORATORY (EE 753/1)
Instruction Sheet for SOFT COMPUTING LABORATORY (EE 753/1) Develop the following programs in the MATLAB environment: 1. Write a program in MATLAB for Feed Forward Neural Network with Back propagation training
More informationSelected Function List from Neural Network Toolbox For Use with MATLAB. User s s Guide Version 4
Selected Function List from Neural Network Toolbox For Use with MATLAB User s s Guide Version 4 Analysis Functions Errsurf: : Error surface of a single input neuron. Maxlinlr: : Maximum learning rate for
More informationMulti-Layer Perceptron in MATLAB NN Toolbox
Multi-Layer Perceptron in MATLAB NN Toolbox [Part 1] Yousof Koohmaskan, Behzad Bahrami, Seyyed Mahdi Akrami, Mahyar AbdeEtedal Department of Electrical Engineering Amirkabir University of Technology (Tehran
More informationIntro to Neural Networks and Deep Learning
Intro to Neural Networks and Deep Learning Jack Lanchantin Dr. Yanjun Qi UVA CS 6316 1 Neurons 1-Layer Neural Network Multi-layer Neural Network Loss Functions Backpropagation Nonlinearity Functions NNs
More informationArtificial Intelligence
Artificial Intelligence Jeff Clune Assistant Professor Evolving Artificial Intelligence Laboratory Announcements Be making progress on your projects! Three Types of Learning Unsupervised Supervised Reinforcement
More informationClassification of the students scores based on some artificial neural networks
Issue 4, Volume 3, 9 7 Classification of the students scores based on some artificial neural networks Hu Hongping Bai Yanping Abstract In this paper, the data of students scores are analyzed by using the
More information4. Multilayer Perceptrons
4. Multilayer Perceptrons This is a supervised error-correction learning algorithm. 1 4.1 Introduction A multilayer feedforward network consists of an input layer, one or more hidden layers, and an output
More informationPattern Classification
Pattern Classification All materials in these slides were taen from Pattern Classification (2nd ed) by R. O. Duda,, P. E. Hart and D. G. Stor, John Wiley & Sons, 2000 with the permission of the authors
More informationShort-term wind forecasting using artificial neural networks (ANNs)
Energy and Sustainability II 197 Short-term wind forecasting using artificial neural networks (ANNs) M. G. De Giorgi, A. Ficarella & M. G. Russo Department of Engineering Innovation, Centro Ricerche Energia
More informationBackpropagation Neural Net
Backpropagation Neural Net As is the case with most neural networks, the aim of Backpropagation is to train the net to achieve a balance between the ability to respond correctly to the input patterns that
More informationCOGS Q250 Fall Homework 7: Learning in Neural Networks Due: 9:00am, Friday 2nd November.
COGS Q250 Fall 2012 Homework 7: Learning in Neural Networks Due: 9:00am, Friday 2nd November. For the first two questions of the homework you will need to understand the learning algorithm using the delta
More informationLecture 10. Neural networks and optimization. Machine Learning and Data Mining November Nando de Freitas UBC. Nonlinear Supervised Learning
Lecture 0 Neural networks and optimization Machine Learning and Data Mining November 2009 UBC Gradient Searching for a good solution can be interpreted as looking for a minimum of some error (loss) function
More information7 Advanced Learning Algorithms for Multilayer Perceptrons
7 Advanced Learning Algorithms for Multilayer Perceptrons The basic back-propagation learning law is a gradient-descent algorithm based on the estimation of the gradient of the instantaneous sum-squared
More informationArtificial Neural Networks Examination, June 2004
Artificial Neural Networks Examination, June 2004 Instructions There are SIXTY questions (worth up to 60 marks). The exam mark (maximum 60) will be added to the mark obtained in the laborations (maximum
More informationSPSS, University of Texas at Arlington. Topics in Machine Learning-EE 5359 Neural Networks
Topics in Machine Learning-EE 5359 Neural Networks 1 The Perceptron Output: A perceptron is a function that maps D-dimensional vectors to real numbers. For notational convenience, we add a zero-th dimension
More informationArtificial Neural Networks. Edward Gatt
Artificial Neural Networks Edward Gatt What are Neural Networks? Models of the brain and nervous system Highly parallel Process information much more like the brain than a serial computer Learning Very
More informationIntroduction to Natural Computation. Lecture 9. Multilayer Perceptrons and Backpropagation. Peter Lewis
Introduction to Natural Computation Lecture 9 Multilayer Perceptrons and Backpropagation Peter Lewis 1 / 25 Overview of the Lecture Why multilayer perceptrons? Some applications of multilayer perceptrons.
More informationNeural Networks, Computation Graphs. CMSC 470 Marine Carpuat
Neural Networks, Computation Graphs CMSC 470 Marine Carpuat Binary Classification with a Multi-layer Perceptron φ A = 1 φ site = 1 φ located = 1 φ Maizuru = 1 φ, = 2 φ in = 1 φ Kyoto = 1 φ priest = 0 φ
More informationNeural network modelling of reinforced concrete beam shear capacity
icccbe 2010 Nottingham University Press Proceedings of the International Conference on Computing in Civil and Building Engineering W Tizani (Editor) Neural network modelling of reinforced concrete beam
More informationPATTERN CLASSIFICATION
PATTERN CLASSIFICATION Second Edition Richard O. Duda Peter E. Hart David G. Stork A Wiley-lnterscience Publication JOHN WILEY & SONS, INC. New York Chichester Weinheim Brisbane Singapore Toronto CONTENTS
More informationSerious limitations of (single-layer) perceptrons: Cannot learn non-linearly separable tasks. Cannot approximate (learn) non-linear functions
BACK-PROPAGATION NETWORKS Serious limitations of (single-layer) perceptrons: Cannot learn non-linearly separable tasks Cannot approximate (learn) non-linear functions Difficult (if not impossible) to design
More informationMachine Learning and Data Mining. Multi-layer Perceptrons & Neural Networks: Basics. Prof. Alexander Ihler
+ Machine Learning and Data Mining Multi-layer Perceptrons & Neural Networks: Basics Prof. Alexander Ihler Linear Classifiers (Perceptrons) Linear Classifiers a linear classifier is a mapping which partitions
More informationUnit III. A Survey of Neural Network Model
Unit III A Survey of Neural Network Model 1 Single Layer Perceptron Perceptron the first adaptive network architecture was invented by Frank Rosenblatt in 1957. It can be used for the classification of
More informationSimple neuron model Components of simple neuron
Outline 1. Simple neuron model 2. Components of artificial neural networks 3. Common activation functions 4. MATLAB representation of neural network. Single neuron model Simple neuron model Components
More informationy(x n, w) t n 2. (1)
Network training: Training a neural network involves determining the weight parameter vector w that minimizes a cost function. Given a training set comprising a set of input vector {x n }, n = 1,...N,
More informationComputational statistics
Computational statistics Lecture 3: Neural networks Thierry Denœux 5 March, 2016 Neural networks A class of learning methods that was developed separately in different fields statistics and artificial
More informationSupporting Information
Supporting Information Convolutional Embedding of Attributed Molecular Graphs for Physical Property Prediction Connor W. Coley a, Regina Barzilay b, William H. Green a, Tommi S. Jaakkola b, Klavs F. Jensen
More informationMultilayer Neural Networks
Multilayer Neural Networks Introduction Goal: Classify objects by learning nonlinearity There are many problems for which linear discriminants are insufficient for minimum error In previous methods, the
More informationLecture 17: Neural Networks and Deep Learning
UVA CS 6316 / CS 4501-004 Machine Learning Fall 2016 Lecture 17: Neural Networks and Deep Learning Jack Lanchantin Dr. Yanjun Qi 1 Neurons 1-Layer Neural Network Multi-layer Neural Network Loss Functions
More informationCSC Neural Networks. Perceptron Learning Rule
CSC 302 1.5 Neural Networks Perceptron Learning Rule 1 Objectives Determining the weight matrix and bias for perceptron networks with many inputs. Explaining what a learning rule is. Developing the perceptron
More informationApplication of Neural Networks for Control of Inverted Pendulum
Application of Neural Networks for Control of Inverted Pendulum VALERI MLADENOV Department of Theoretical Electrical Engineering Technical University of Sofia Sofia, Kliment Ohridski blvd. 8; BULARIA valerim@tu-sofia.bg
More informationModelling of Pehlivan-Uyaroglu_2010 Chaotic System via Feed Forward Neural Network and Recurrent Neural Networks
Modelling of Pehlivan-Uyaroglu_2010 Chaotic System via Feed Forward Neural Network and Recurrent Neural Networks 1 Murat ALÇIN, 2 İhsan PEHLİVAN and 3 İsmail KOYUNCU 1 Department of Electric -Energy, Porsuk
More informationIntroduction to Neural Networks
Introduction to Neural Networks What are (Artificial) Neural Networks? Models of the brain and nervous system Highly parallel Process information much more like the brain than a serial computer Learning
More informationECLT 5810 Classification Neural Networks. Reference: Data Mining: Concepts and Techniques By J. Hand, M. Kamber, and J. Pei, Morgan Kaufmann
ECLT 5810 Classification Neural Networks Reference: Data Mining: Concepts and Techniques By J. Hand, M. Kamber, and J. Pei, Morgan Kaufmann Neural Networks A neural network is a set of connected input/output
More informationIntroduction to Neural Networks
CUONG TUAN NGUYEN SEIJI HOTTA MASAKI NAKAGAWA Tokyo University of Agriculture and Technology Copyright by Nguyen, Hotta and Nakagawa 1 Pattern classification Which category of an input? Example: Character
More informationNeuro-Fuzzy Comp. Ch. 4 March 24, R p
4 Feedforward Multilayer Neural Networks part I Feedforward multilayer neural networks (introduced in sec 17) with supervised error correcting learning are used to approximate (synthesise) a non-linear
More informationA STATE-SPACE NEURAL NETWORK FOR MODELING DYNAMICAL NONLINEAR SYSTEMS
A STATE-SPACE NEURAL NETWORK FOR MODELING DYNAMICAL NONLINEAR SYSTEMS Karima Amoura Patrice Wira and Said Djennoune Laboratoire CCSP Université Mouloud Mammeri Tizi Ouzou Algeria Laboratoire MIPS Université
More informationKeywords- Source coding, Huffman encoding, Artificial neural network, Multilayer perceptron, Backpropagation algorithm
Volume 4, Issue 5, May 2014 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Huffman Encoding
More informationLearning strategies for neuronal nets - the backpropagation algorithm
Learning strategies for neuronal nets - the backpropagation algorithm In contrast to the NNs with thresholds we handled until now NNs are the NNs with non-linear activation functions f(x). The most common
More informationCSC 578 Neural Networks and Deep Learning
CSC 578 Neural Networks and Deep Learning Fall 2018/19 3. Improving Neural Networks (Some figures adapted from NNDL book) 1 Various Approaches to Improve Neural Networks 1. Cost functions Quadratic Cross
More informationArtificial Neural Network II MATLAB Neural Network Toolbox
Artificial Neural Network II MATLAB Neural Network Toolbox Werapon Chiracharit Department of Electronic and Telecommunication Engineering King Mongkut s University of Technology Thonburi Input Layer Hidden
More informationAI Programming CS F-20 Neural Networks
AI Programming CS662-2008F-20 Neural Networks David Galles Department of Computer Science University of San Francisco 20-0: Symbolic AI Most of this class has been focused on Symbolic AI Focus or symbols
More informationESTIMATION OF HOURLY MEAN AMBIENT TEMPERATURES WITH ARTIFICIAL NEURAL NETWORKS 1. INTRODUCTION
Mathematical and Computational Applications, Vol. 11, No. 3, pp. 215-224, 2006. Association for Scientific Research ESTIMATION OF HOURLY MEAN AMBIENT TEMPERATURES WITH ARTIFICIAL NEURAL NETWORKS Ömer Altan
More informationMultilayer Perceptron = FeedForward Neural Network
Multilayer Perceptron = FeedForward Neural Networ History Definition Classification = feedforward operation Learning = bacpropagation = local optimization in the space of weights Pattern Classification
More informationArtifical Neural Networks
Neural Networks Artifical Neural Networks Neural Networks Biological Neural Networks.................................. Artificial Neural Networks................................... 3 ANN Structure...........................................
More informationDeep Learning & Artificial Intelligence WS 2018/2019
Deep Learning & Artificial Intelligence WS 2018/2019 Linear Regression Model Model Error Function: Squared Error Has no special meaning except it makes gradients look nicer Prediction Ground truth / target
More informationResearch Article Calculation for Primary Combustion Characteristics of Boron-Based Fuel-Rich Propellant Based on BP Neural Network
Combustion Volume 2012, Article ID 635190, 6 pages doi:10.1155/2012/635190 Research Article Calculation for Primary Combustion Characteristics of Boron-Based Fuel-Rich Propellant Based on BP Neural Network
More informationCSC242: Intro to AI. Lecture 21
CSC242: Intro to AI Lecture 21 Administrivia Project 4 (homeworks 18 & 19) due Mon Apr 16 11:59PM Posters Apr 24 and 26 You need an idea! You need to present it nicely on 2-wide by 4-high landscape pages
More informationNONLINEAR CLASSIFICATION AND REGRESSION. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition
NONLINEAR CLASSIFICATION AND REGRESSION Nonlinear Classification and Regression: Outline 2 Multi-Layer Perceptrons The Back-Propagation Learning Algorithm Generalized Linear Models Radial Basis Function
More informationNeural Networks and Fuzzy Logic Rajendra Dept.of CSE ASCET
Unit-. Definition Neural network is a massively parallel distributed processing system, made of highly inter-connected neural computing elements that have the ability to learn and thereby acquire knowledge
More informationIntroduction to Neural Networks: Structure and Training
Introduction to Neural Networks: Structure and Training Professor Q.J. Zhang Department of Electronics Carleton University, Ottawa, Canada www.doe.carleton.ca/~qjz, qjz@doe.carleton.ca A Quick Illustration
More informationResearch Article Predicting Global Solar Radiation Using an Artificial Neural Network Single-Parameter Model
Artificial Neural Systems Volume 2011, Article ID 751908, 7 pages doi:10.1155/2011/751908 Research Article Predicting Global Solar Radiation Using an Artificial Neural Network Single-Parameter Model Karoro
More informationBased on the original slides of Hung-yi Lee
Based on the original slides of Hung-yi Lee Google Trends Deep learning obtains many exciting results. Can contribute to new Smart Services in the Context of the Internet of Things (IoT). IoT Services
More informationStatistical Machine Learning from Data
January 17, 2006 Samy Bengio Statistical Machine Learning from Data 1 Statistical Machine Learning from Data Multi-Layer Perceptrons Samy Bengio IDIAP Research Institute, Martigny, Switzerland, and Ecole
More informationSimple Neural Nets For Pattern Classification
CHAPTER 2 Simple Neural Nets For Pattern Classification Neural Networks General Discussion One of the simplest tasks that neural nets can be trained to perform is pattern classification. In pattern classification
More informationEPL442: Computational
EPL442: Computational Learning Systems Lab 2 Vassilis Vassiliades Department of Computer Science University of Cyprus Outline Artificial Neuron Feedforward Neural Network Back-propagation Algorithm Notes
More informationLecture 7 Artificial neural networks: Supervised learning
Lecture 7 Artificial neural networks: Supervised learning Introduction, or how the brain works The neuron as a simple computing element The perceptron Multilayer neural networks Accelerated learning in
More informationHopfield Network Recurrent Netorks
Hopfield Network Recurrent Netorks w 2 w n E.P. P.E. y (k) Auto-Associative Memory: Given an initial n bit pattern returns the closest stored (associated) pattern. No P.E. self-feedback! w 2 w n2 E.P.2
More informationTitle: Artificial Neural Network and Fuzzy Logic in forecasting short-term Temperature
Case project report 2013 Title: Artificial Neural Network and Fuzzy Logic in forecasting short-term Temperature Candidates: Chao Xi (120295) Erihe (120293) Telemark University College Faculty of Technology
More informationPattern Recognition Prof. P. S. Sastry Department of Electronics and Communication Engineering Indian Institute of Science, Bangalore
Pattern Recognition Prof. P. S. Sastry Department of Electronics and Communication Engineering Indian Institute of Science, Bangalore Lecture - 27 Multilayer Feedforward Neural networks with Sigmoidal
More informationStudy on the use of neural networks in control systems
Study on the use of neural networks in control systems F. Rinaldi Politecnico di Torino & Italian Air Force, Italy Keywords: neural networks, automatic control Abstract The purpose of this report is to
More informationNeural networks. Chapter 19, Sections 1 5 1
Neural networks Chapter 19, Sections 1 5 Chapter 19, Sections 1 5 1 Outline Brains Neural networks Perceptrons Multilayer perceptrons Applications of neural networks Chapter 19, Sections 1 5 2 Brains 10
More informationData Mining Part 5. Prediction
Data Mining Part 5. Prediction 5.5. Spring 2010 Instructor: Dr. Masoud Yaghini Outline How the Brain Works Artificial Neural Networks Simple Computing Elements Feed-Forward Networks Perceptrons (Single-layer,
More informationCS60010: Deep Learning
CS60010: Deep Learning Sudeshna Sarkar Spring 2018 16 Jan 2018 FFN Goal: Approximate some unknown ideal function f : X! Y Ideal classifier: y = f*(x) with x and category y Feedforward Network: Define parametric
More informationApprentissage, réseaux de neurones et modèles graphiques (RCP209) Neural Networks and Deep Learning
Apprentissage, réseaux de neurones et modèles graphiques (RCP209) Neural Networks and Deep Learning Nicolas Thome Prenom.Nom@cnam.fr http://cedric.cnam.fr/vertigo/cours/ml2/ Département Informatique Conservatoire
More informationLecture 5: Logistic Regression. Neural Networks
Lecture 5: Logistic Regression. Neural Networks Logistic regression Comparison with generative models Feed-forward neural networks Backpropagation Tricks for training neural networks COMP-652, Lecture
More informationStability of backpropagation learning rule
Stability of backpropagation learning rule Petr Krupanský, Petr Pivoňka, Jiří Dohnal Department of Control and Instrumentation Brno University of Technology Božetěchova 2, 612 66 Brno, Czech republic krupan,
More informationComputational Intelligence Winter Term 2017/18
Computational Intelligence Winter Term 207/8 Prof. Dr. Günter Rudolph Lehrstuhl für Algorithm Engineering (LS ) Fakultät für Informatik TU Dortmund Plan for Today Single-Layer Perceptron Accelerated Learning
More information<Special Topics in VLSI> Learning for Deep Neural Networks (Back-propagation)
Learning for Deep Neural Networks (Back-propagation) Outline Summary of Previous Standford Lecture Universal Approximation Theorem Inference vs Training Gradient Descent Back-Propagation
More informationCOMPUTATIONAL INTELLIGENCE (INTRODUCTION TO MACHINE LEARNING) SS16
COMPUTATIONAL INTELLIGENCE (INTRODUCTION TO MACHINE LEARNING) SS6 Lecture 3: Classification with Logistic Regression Advanced optimization techniques Underfitting & Overfitting Model selection (Training-
More informationNeural Network Control of an Inverted Pendulum on a Cart
Neural Network Control of an Inverted Pendulum on a Cart VALERI MLADENOV, GEORGI TSENOV, LAMBROS EKONOMOU, NICHOLAS HARKIOLAKIS, PANAGIOTIS KARAMPELAS Department of Theoretical Electrical Engineering Technical
More informationEKF LEARNING FOR FEEDFORWARD NEURAL NETWORKS
EKF LEARNING FOR FEEDFORWARD NEURAL NETWORKS A. Alessandri, G. Cirimele, M. Cuneo, S. Pagnan, M. Sanguineti Institute of Intelligent Systems for Automation, ISSIA-CNR National Research Council of Italy,
More informationSupervised (BPL) verses Hybrid (RBF) Learning. By: Shahed Shahir
Supervised (BPL) verses Hybrid (RBF) Learning By: Shahed Shahir 1 Outline I. Introduction II. Supervised Learning III. Hybrid Learning IV. BPL Verses RBF V. Supervised verses Hybrid learning VI. Conclusion
More informationARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD
ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD WHAT IS A NEURAL NETWORK? The simplest definition of a neural network, more properly referred to as an 'artificial' neural network (ANN), is provided
More informationNeural Networks DWML, /25
DWML, 2007 /25 Neural networks: Biological and artificial Consider humans: Neuron switching time 0.00 second Number of neurons 0 0 Connections per neuron 0 4-0 5 Scene recognition time 0. sec 00 inference
More informationArtificial Neural Network : Training
Artificial Neural Networ : Training Debasis Samanta IIT Kharagpur debasis.samanta.iitgp@gmail.com 06.04.2018 Debasis Samanta (IIT Kharagpur) Soft Computing Applications 06.04.2018 1 / 49 Learning of neural
More informationVorlesung Neuronale Netze - Radiale-Basisfunktionen (RBF)-Netze -
Vorlesung Neuronale Netze - Radiale-Basisfunktionen (RBF)-Netze - SS 004 Holger Fröhlich (abg. Vorl. von S. Kaushik¹) Lehrstuhl Rechnerarchitektur, Prof. Dr. A. Zell ¹www.cse.iitd.ernet.in/~saroj Radial
More informationComputational Intelligence
Plan for Today Single-Layer Perceptron Computational Intelligence Winter Term 00/ Prof. Dr. Günter Rudolph Lehrstuhl für Algorithm Engineering (LS ) Fakultät für Informatik TU Dortmund Accelerated Learning
More informationNeural Network Toolbox. Version (R11) 01-Jul Adapt functions. adaptwb - By-weight-and-bias network adaption function.
Neural Network Toolbox. Version 3.0.1 (R11) 01-Jul-1998 Adapt functions. adaptwb - By-weight-and-bias network adaption function. Analysis functions. errsurf - Error surface of single input neuron. maxlinlr
More informationArtificial Neural Networks
Artificial Neural Networks 鮑興國 Ph.D. National Taiwan University of Science and Technology Outline Perceptrons Gradient descent Multi-layer networks Backpropagation Hidden layer representations Examples
More informationNeural Networks with Applications to Vision and Language. Feedforward Networks. Marco Kuhlmann
Neural Networks with Applications to Vision and Language Feedforward Networks Marco Kuhlmann Feedforward networks Linear separability x 2 x 2 0 1 0 1 0 0 x 1 1 0 x 1 linearly separable not linearly separable
More informationMultilayer Perceptrons and Backpropagation
Multilayer Perceptrons and Backpropagation Informatics 1 CG: Lecture 7 Chris Lucas School of Informatics University of Edinburgh January 31, 2017 (Slides adapted from Mirella Lapata s.) 1 / 33 Reading:
More informationMachine Learning. Neural Networks. (slides from Domingos, Pardo, others)
Machine Learning Neural Networks (slides from Domingos, Pardo, others) Human Brain Neurons Input-Output Transformation Input Spikes Output Spike Spike (= a brief pulse) (Excitatory Post-Synaptic Potential)
More informationArtificial Neural Networks
Artificial Neural Networks Threshold units Gradient descent Multilayer networks Backpropagation Hidden layer representations Example: Face Recognition Advanced topics 1 Connectionist Models Consider humans:
More informationCourse 395: Machine Learning - Lectures
Course 395: Machine Learning - Lectures Lecture 1-2: Concept Learning (M. Pantic) Lecture 3-4: Decision Trees & CBC Intro (M. Pantic & S. Petridis) Lecture 5-6: Evaluating Hypotheses (S. Petridis) Lecture
More informationNeural Networks. Chapter 18, Section 7. TB Artificial Intelligence. Slides from AIMA 1/ 21
Neural Networks Chapter 8, Section 7 TB Artificial Intelligence Slides from AIMA http://aima.cs.berkeley.edu / 2 Outline Brains Neural networks Perceptrons Multilayer perceptrons Applications of neural
More informationArtificial Neural Networks Examination, March 2004
Artificial Neural Networks Examination, March 2004 Instructions There are SIXTY questions (worth up to 60 marks). The exam mark (maximum 60) will be added to the mark obtained in the laborations (maximum
More informationMachine Learning for Large-Scale Data Analysis and Decision Making A. Neural Networks Week #6
Machine Learning for Large-Scale Data Analysis and Decision Making 80-629-17A Neural Networks Week #6 Today Neural Networks A. Modeling B. Fitting C. Deep neural networks Today s material is (adapted)
More informationThree Structures of a Multilayer Artificial Neural Network for Predicting the Solar Radiation of Baghdad City- Iraq
Three Structures of a Multilayer Artificial Neural Network for Predicting the Solar Radiation of Baghdad City- Iraq a,* Abdulrahman Th. Mohammad, b khalil I. Mahmood, c Dawood S. Mahjoob a,b,c Baqubah
More informationNeural Networks Based on Competition
Neural Networks Based on Competition In some examples of pattern classification we encountered a situation in which the net was trained to classify the input signal into one of the output categories, while
More informationMultilayer Neural Networks
Multilayer Neural Networks Multilayer Neural Networks Discriminant function flexibility NON-Linear But with sets of linear parameters at each layer Provably general function approximators for sufficient
More informationProject 1: A comparison of time delay neural networks (TDNN) trained with mean squared error (MSE) and error entropy criterion (EEC)
1 Project 1: A comparison of time delay neural networks (TDNN) trained with mean squared error (MSE) and error entropy criterion (EEC) Stefan Craciun Abstract The goal is to implement a TDNN (time delay
More informationA Novel Activity Detection Method
A Novel Activity Detection Method Gismy George P.G. Student, Department of ECE, Ilahia College of,muvattupuzha, Kerala, India ABSTRACT: This paper presents an approach for activity state recognition of
More informationAdvanced statistical methods for data analysis Lecture 2
Advanced statistical methods for data analysis Lecture 2 RHUL Physics www.pp.rhul.ac.uk/~cowan Universität Mainz Klausurtagung des GK Eichtheorien exp. Tests... Bullay/Mosel 15 17 September, 2008 1 Outline
More information1 What a Neural Network Computes
Neural Networks 1 What a Neural Network Computes To begin with, we will discuss fully connected feed-forward neural networks, also known as multilayer perceptrons. A feedforward neural network consists
More informationCSE 352 (AI) LECTURE NOTES Professor Anita Wasilewska. NEURAL NETWORKS Learning
CSE 352 (AI) LECTURE NOTES Professor Anita Wasilewska NEURAL NETWORKS Learning Neural Networks Classifier Short Presentation INPUT: classification data, i.e. it contains an classification (class) attribute.
More informationLECTURE # - NEURAL COMPUTATION, Feb 04, Linear Regression. x 1 θ 1 output... θ M x M. Assumes a functional form
LECTURE # - EURAL COPUTATIO, Feb 4, 4 Linear Regression Assumes a functional form f (, θ) = θ θ θ K θ (Eq) where = (,, ) are the attributes and θ = (θ, θ, θ ) are the function parameters Eample: f (, θ)
More informationCS 6501: Deep Learning for Computer Graphics. Basics of Neural Networks. Connelly Barnes
CS 6501: Deep Learning for Computer Graphics Basics of Neural Networks Connelly Barnes Overview Simple neural networks Perceptron Feedforward neural networks Multilayer perceptron and properties Autoencoders
More informationEEE 241: Linear Systems
EEE 4: Linear Systems Summary # 3: Introduction to artificial neural networks DISTRIBUTED REPRESENTATION An ANN consists of simple processing units communicating with each other. The basic elements of
More informationRevision: Neural Network
Revision: Neural Network Exercise 1 Tell whether each of the following statements is true or false by checking the appropriate box. Statement True False a) A perceptron is guaranteed to perfectly learn
More information