COMPARISON STUDY OF SENSITIVITY DEFINITIONS OF NEURAL NETWORKS
|
|
- Thomas Neal
- 6 years ago
- Views:
Transcription
1 COMPARISON SUDY OF SENSIIVIY DEFINIIONS OF NEURAL NEORKS CHUN-GUO LI 1, HAI-FENG LI 1 Machine Learning Center, Facult of Mathematics and Computer Science, Hebei Universit, Baoding 07100, China Department of educational administration, Hebei Universit, Baoding 07100, China licg@hbu.cn, lihf@hbu.cn Abstract: his paper compares the sensitivit definitions of neural networs output to input and weight perturbations. Based on the essence of the sensitivit definitions, the authors classif these sensitivit definitions into 3 categories: Noise-to- Signal Ratio, Geometrical propert of derivative, Angle perturbation in Geometr space. he characteristics of these 3 categories of sensitivit definition are discussed respectivel. It is sensible to classif these sensitivit definitions based on the essence of them for researchers can find other new sensitivit definitions of neural networs. Kewords: Sensitivit definition; categories; neural networs; comparison 1. Introduction Sensitivit analsis is an important issue of neural networs research. here had been large amount of sensitivit definitions of different neural networs emerged in the past decades [1-15]. Sensitivit analsis of neural networ s output to input and weight perturbation is expected to settle some appling problems [16-19]. It can be used to predict the effect of perturbations on the networ s output, quantif the error-tolerance and generalization abilities, guide designing a robust networ, and so on. herefore, there are more and more researchers mae much effort to do the sensitivit analsis of neural networ. hese proposed sensitivit definitions are based on distinguished methodologies and on different neural networs. In one hand, some sensitivit definitions are raised to appl on an ensemble of neural networs, while some on one special neural networ. Man sensitivit definitions are brought forward. For one special neural networ (eg. Multilaer perceptron), It is necessar to find some wa to classif these definitions. Consequentl, people can adopt different definition to settle down different problem to get the best results. here are man was to classif the sensitivit definitions. Some people thought that the sensitivit is defined in two approaches [5]: analtical approach and statistical approach. his classification is based on the measure space of the sensitivit definition. Another wa to classif the sensitivit definitions is based on the neural networs that the definition is applied to. Some definitions can be applied to MLP while some limited on Madalines. Some sensitivit definitions can be used for linear neural networs while some for nonlinear neural networs. hese was to classif the sensitivit definitions did not distinguish the essences of those definitions. In this paper, we studied the sensitivit definitions of the neural networs output to input and weight perturbations. According to the essence of the sensitivit definitions, we classified the sensitivit definitions into 3 categories: Noise-to- Signal Ratio, Geometrical propert of derivative, Angle perturbation in Geometr space. Some people proposed the ideas separatel, and some others defined the sensitivit definitions following and improving the idea. his would be sensible for people to find other new sensitivit definitions of neural networs.. Categories of sensitivit definition of neural networs Since the sensitivit of the neural networs emerged, man ideas are proposed to define the sensitivit of neural networs output to input and weight perturbations. Some of these ideas improved the original ideas, while some of them are applied to practical problems. he ideas were reformulated before application. Subsequentl, man new expression formulae appeared. Gathered these formulae together and detected their essences, we classif these sensitivit definitions into 3 approaches according to the borrowed ideas. e describe them in detail in the following 3 sections..1. Noise-to-Signal Ratio
2 Piché [1] raised the idea of Noise-to-Signal for the selection of weight accuracies for Madalines. his idea of sensitivit definition is based on the ratio of noise and signal. Noise, produced from the environment of the signal, affects the transformation of the signal. he Noise-to-Signal Ratio qualifies how much the noise affects the transformation of signal. Based on this idea, Piché regarded the noise to the signal as the perturbations to inputs and weights of the neural networs. He too into account the mathematical characteristics of random variables and proposed the sensitivit definition of neural networs. He defined the sensitivit definition on an Adaline and extended to Madaline. Yeung et al extended Piché s wor on Madalines to MLP and RBFNN [-8]. Given a structured neural networ, all the parameters of the networ are considered as random variables. If the output of the networ can be expressed b, the sensitivit of the networ to input and weight perturbations could be defined as: where NSR (1) is the variance of the output errors caused b weight or first laer input perturbations, and is the variance of the output. Piché derived the calculating formulae of Eq.(1) of Madaline for the input and weight sensitivit in the condition of the following issues: 1. It is assumed that an ensemble of possible ideal neural networs can be identified. hese ideal networs ma represent a general set of Madalines.. he architecture of all ideal neural networs is assumed to be identical. he networ has L laers with N l neurons full connected to the next laer. here is no bias input in the networ. he output of the networ is nonlinear, and all output operators are odd functions. 3. All the weights associated with each node of the networ are modeled as a vector of random variables. Over the ensemble of networs, the weights of a particular laer all have the same variance. he mean value of each weight of the networ over the ensemble is zero. All in all, the components of vector are considered to be zero-mean independent identicall distributed (iid) random variables. he distribution of these random variables is not particularl important. 4. he inputs of the networ are also modeled as a vector of random variables X. he inputs of the networ are also zero-mean independent identicall distributed (iid) random variables. he distribution of these random variables is not particularl important. (Combined the above points 3 and 4, the components of the input and output vector of each laer are independent identicall distributed (iid) random variables.) 5. he mathematical expectation of noise is assumed to be E ln 0 zero. Yeung et al extended Piché s wor on Madalines to MLP and RBFNN [-8]. Although the calculating formulae are reformulated, the sensitivit definition is based on the noise and signal ratio. he differences between them are the procedure to derive the expressions of the numerator and denominator of Noise-to-Signal Ratio. For example, ing et al [] extended the sensitivit definition from Madaline to RBFNN. he considered the sensitivit definition of neural networs output to input perturbations. he relative sensitivit is defined as the following formulae: D X X X SRD () D X where X is the input vector of the neural networ, X the perturbation of the input, D the variance of, the -th output of the networ. ing et al educed the calculating formulae of Eq.() b the calculating properties of variance for RBFNN. Yeung et al [7] reformulated the sensitivit definition and considered Multilaer perceptron with antismmetric squashing activation function. he introduced both the input and weight perturbation of the neural networ. he sensitivit is defined as the following formula: S RD Dg X D g X X D g X where g is the function expression of the networ. Eqs.(-3) were relative sensitivit as the said. Actuall, the two sensitivit formulae are defined in the same wa. he are the expression formulae of Eq.(1). Although these definitions were defined based on the same idea as Eq.(1), the other authors improved the formulae in different expression. For example, beside relative sensitivit definitions, ing et al [,7] also defined the absolute sensitivit definition without the denominator in Eq.(-3), shown as the following two formulae: S D X X X (4) D D (3) S D g X X g X (5) he above absolute sensitivit definition is defined according to the meaning of the variance as mathematical characteristic of random variable. Variance measures the expected deviation amplitude of the value of the random variable from the expected value of the random variable.
3 .. Geometrical propert of derivative First, let us remind some definition in analtical mathematics. Given a differentiable continuous function f x, its first moment can be derived to be f x with respect to x. f x 0 is the slope of function f x in the point x 0, denoting the change rate around x 0. Correspondingl, f x is the function of slope in the definition field of x. Based on this meaning, Zurada et al [9] proposed that the sensitivit of neural networs output to input perturbation can be defined as the derivative of the neural networs. Zurada et al considered a MFNN with a single hidden laer. he networ is regarded as a nonlinear and f x, where differentiable function,,,,,,, 1 x x1 x x n, and n are the number of output and input neurons respectivel. Since is differentiable at x, Jocobian matrix Eq.(6) is derived firstl. he matrix can be evaluated over the entire training set according to several was such as mean square average, absolute value average, and maximum value. Following this sensitivit definition, similarl definitions are proposed to improve and extend the sensitivit definition. ang et al [8] extended this definition to random measure space and defined the sensitivit of RBFNN s output to input regarding the input vector as random vector x1 x x n J x1 x xn (6) x1 x x n he characteristics of this sensitivit definition is listed as the following issues: 1. he neural networs are required to be differentiable and feedforward. his sensitivit definition is unsuitable for perceptron. Because the active function of a perceptron is non-differentiable.. he sensitivit is defined for the neural networs with continuous valued input vector. If the input vector is random vector, this sensitivit definitions would be modified, just as in [8]. 3. Zurada et al defined the sensitivit of the networs outputs to input perturbations. It can be extend to the sensitivit to weight perturbations if the networ is differentiable with respect to the weight variables. 4. his definition sufficientl maes use of the training set to calculate the sensitivit of each input variable. 5. he sensitivit of the input of the networ is calculated for trained networ. herefore, the training algorithm of the networ affects the absolute sensitivit of the input..3. Angle perturbation in Geometr space Marhelen Stevenson et al proposed this idea to define the sensitivit in the geometr space. he too the input and weight perturbation as errors to input and weight. he defined the sensitivit as the probabilit of decision error due to input and weight error. he sensitivit definition is extended from Adaline to Madaline. For an Adaline with n+1 variable inputs X x0 x1 x x n including a threshold input:,,,,, the inputs tae on binar values of either +1 or 1. Associated with the Adaline are n+1 adjustable analog weights: w0, w1, w,, w n. he input-output map of the Adaline can be summarized as 1, 0 output of Adaline X (7) 1, X 0 A Madaline is a laered networ of Adaline elements. herefore the sensitivit definition of Adaline can be extended to Madaline. Marhelen Stevenson et al firstl defined the sensitivit of Adaline. Assuming binar-valued inputs, there are n possible input patterns for an Adaline with n variable inputs. Each input pattern corresponds to a point in n space, which lies on a hpersphere of radius n centered at the origin. here is a hperplane perpendicular to which passes through the origin, where 0 1 x, x, x,, x : X 0. his hperplane, w n, separates the X for which X 0 from those X for which X 0 (i.e. it separates those input vectors which ield a +1 response from those input vectors which ield a 1 response). divides the hpersphere into two hemi- hperspheres. If a randoml oriented perturbation vector is added to, the perturbed weight vector corresponds to another hperplane p p. p would probabl produce decision errors. A decision error is made when an arbitrar input vector is mapped into opposite output categories b and p. he probabilit of a decision error due to an weight perturbation is defined:
4 P Decision Error / (8) where is the angular between and p. It is the sensitivit definition to weight errors. For practical applications, the probabilit of a decision error is derived in terms of the weight perturbation ratio. his definition is based on the Hoff hpersphere-area approximation. he Hoff hpersphere-area approximation states that as n gets large, the points corresponding to the n-dimensional input patterns are approximatel uniforml distributed over the surface of a hpersphere in n space. Consequentl, the percentage of input patterns, which correspond to points on a selected region of the hpersphere, can be approximated as the ratio of the surface content of the selected region to the surface content of the entire hpersphere. his sensitivit definition is convenienc to calculate the sensitivit to weight perturbations. he calculating formulae do not depend on the structure of the networ. hat the sensitivit utilizes is the geometr distribution of the training patterns. he characteristics of this sense of sensitivit definition is listed as the following points: 1. he input vectors are required to be binar-valued, and the output active function of the networ is required to be hard limited function. he active function is non-differentiable. Reminding the limitation of the sensitivit definition, the sensitivit definition here is an alternative definition when the active function of the networ is non-differentiable.. hen n is small (eg. n=10), the points corresponding to the n-dimensional input patterns are hardl possible uniforml distributed over the surface of a hpersphere in n space. In this condition, the calculated probabilit b Eq.(8) would be hard to believe. 3. his stud focused on the sensitivit of Adaline and Madaline output to weight error. For the sensitivit of Adaline and Madaline output to input error, it leaves as the further research, or taes the other wa to define the sensitivit. And for the other neural networs, we should thin over whether this wa to define the sensitivit is suitable or not. 4. he approximating computation formulae are just for large enough n and small enough. Cheng and Yeung [14] also used a geometrical technique to investigate the sensitivit of Neocognitions b modifing the hpersphere model. Unfortunatel the hpersphere model the used fails for the case of the MLP because its input and weight vectors generall do not span a hpersphere surface or volume. 3. Comparison of sensitivit categories he above 3 categories of the sensitivit definitions are modeled based on distinguished methodologies. here are advantages and disadvantages when these definitions are used in applications. he table below gives a straight sight into the comparison of these 3 categories of the sensitivit definitions. Methodolog Noise-to-Signal Ratio Geometr propert of derivative Angle perturbation able 1. Comparison of sensitivit definitions of different methodologies. Input or weight Ensemble of neural perturbation Requirement of Input vector networ the networ Input eight Madaline Adaline Multilaer Madaline Adaline Input and output vector are iid respectivel. Differentiable Binar valued input and output Random Continuous valued Discrete valued From the table above, we can see clearl that different sensitivit definitions are suitable to deal with different ensembles of networs. In application, mabe of the 3 sensitivit definitions are suitable to define the sensitivit definitions of the networs output to input perturbations or weight perturbations. For example, if we are studing the sensitivit of the Adaline s output to weight perturbations, both Noise-to-Signal ratio and angle perturbation can be used to help to sensitivit analsis. But which of them is better to sensitivit analsis is left to further research. 4. Conclusion his paper compares several sensitivit definitions of
5 neural networs output to input and weight perturbations. he sensitivit definitions are classified into 3 categories according to the essence of the sensitivit definition. hese 3 categories of sensitivit definitions can deal with different ensemble of neural networs. his wa of classif the sensitivit classification can help to detect the new sensitivit definition of neural networs output to input and weight perturbations. here are still man points to focus on. From able 1, we can see that it is suitable to define the sensitivit definition of a feedforward neural networ s output to weight perturbation based on all the 3 categories. So, which is the best for the sensitivit analsis in this situation is left for the further research. It is necessar to carr out plent of experiments to help discover what is the best answer of the above problem. References [1] Stephen. Piché: he Selection of eight Accuracies for Madalines. IEEE ransactions on Neural Networs. Vol.6, No., March 1995, [] ing.y. Ng, Daniel S. Yeung, Q. Ran, Eric C.C. sang: Statistical Output Sensitivit to Input and eight Perturbations of Radial Basis Function Neuron Networs. IEEE Proceedings of the International Conferences on Sstem, Man and Cbernetics, 00, [3] Jin Young Choi, Chong-Ho Choi: Sensitivit Analsis of Multilaer Perceptron with Differentiable Activation Functions. IEEE ransactions on Neural Networs. Vol.3, No.1, Januar 199, [4] ing.y. Ng, D.S. Yeung: Electronics Letters, 15 th Ma 003, Vol.39, No. 10, [5] Xiaoqin Zeng, Daniel S. Yeung: Sensitivit Analsis of Multilaer Perceptron to Input and eight Perturbations. IEEE ransactions on Neural Networs. Vol.1, No.6, November 001, [6] Xiaoqin Zeng, Daniel S. Yeung: A Quantified Sensitivit Measure for Multilaer Perceptron to Input Perturbation. Nural Computation, 15, 003, [7] Daniel S. Yeung, Xuequan Sun: Using Function Approximation to Analze the Sensitivit of MLP with Antismmetric Squashing Activation Function. IEEE ransactions on Neural Networs, Vol. 13, No. 1, Januar 00, [8] Xizhao ang, Chunguo Li: A New Definition of Sensitivit for RBFNN and Its Applications to Feature Reduction. Second International Smposium on Neural Networs, LNCS 3496, 005, pp [9] Jace M. Zurada, Alesander Malinowsi, Shiro Usui. Perturbation method for deleting redundant inputs of perceptron networs. Neurocomputing 14 (1997) [10] Sherif Hashem: Sensitivit Analsis for Feedforward Artificial Neural Networs with Differentiable Activation Functions. Proceedings of the 199 International Joint Conference on Neural Networs, Baltimore, IEEE, New Jerse, Vol. I, pp [11] Masato Koda: Neural Networ Learning Based on Stochastic Sensitivit Analsis. IEEE ransactions on Sstems, Man and Cbernetics-Part B: Cbernetics. Vol. 7, No.1, Februar 1997, [1] Lale Ozilmaz, ula Yildirim: Sensitivit Analsis for Conic Section Function Neural Networs. IJCNN (4) 000,44-5 [13] Marhelen Stevenson, Rodne inter, Bernard idrow: Sensitivit of Feedforward Neural Networs to eight Errors. IEEE ransactions on Neural Networs, Vol. 1, No. 1, March 1990, [14] Anton Y. Cheng, Daniel S. Yeung: Sensitivit Analsis of Neocognitron. IEEE ransactions on Sstems, Man and Cbernetics-Part C: Applications and Reviews, Vol. 9, No., Ma 1999, [15] Cesare Alippi, Vincenzo Piuri, Mariagiovanna Sami: Sensitivit to Errors in Artificial Neural Networs: A Behavioral Approach. IEEE ransactions on Circuits and Sstems-I: Fundamental heor and Applications, Vol. 4, No. 6, June 1995, [16] D. Shi, D.S. Yeung, J. Gao. Sensitivit analsis applied to the construction of radial basis function networs. Neural Networs, 18 (005), [17] ing.y. Ng, Daniel S. Yeung. Input dimensionalit reduction for Radial Basis Neural Networ classification problems using sensitivit analsis. Proceedings of the First International Conference on Machine Learning and Cbernetics, Beijing, 00, [18] ing.y. Ng, Roc K.C. Chang, Daniel S. Yeung: Dimensionalit Reduction for Denial of Service Detection Problems Using RBFNN Output Sensitivit. Proceedings of the Second International Conference on Machine Learning and Cbernetics, Xi an, 003, [19] Nicolaos B. Karaiannis: Reformulated Radial Basis Neural Networs rained b Gradient Decent. IEEE ransactions on Neural Networs, Vol. 10, No. 3, Ma 1999,
Simple Neural Nets For Pattern Classification
CHAPTER 2 Simple Neural Nets For Pattern Classification Neural Networks General Discussion One of the simplest tasks that neural nets can be trained to perform is pattern classification. In pattern classification
More informationMultilayer Neural Networks
Pattern Recognition Multilaer Neural Networs Lecture 4 Prof. Daniel Yeung School of Computer Science and Engineering South China Universit of Technolog Outline Introduction (6.) Artificial Neural Networ
More informationNeural Networks and the Back-propagation Algorithm
Neural Networks and the Back-propagation Algorithm Francisco S. Melo In these notes, we provide a brief overview of the main concepts concerning neural networks and the back-propagation algorithm. We closely
More informationMultilayer Perceptron
Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Single Perceptron 3 Boolean Function Learning 4
More informationArtificial Neural Networks The Introduction
Artificial Neural Networks The Introduction 01001110 01100101 01110101 01110010 01101111 01101110 01101111 01110110 01100001 00100000 01110011 01101011 01110101 01110000 01101001 01101110 01100001 00100000
More informationUnit III. A Survey of Neural Network Model
Unit III A Survey of Neural Network Model 1 Single Layer Perceptron Perceptron the first adaptive network architecture was invented by Frank Rosenblatt in 1957. It can be used for the classification of
More informationARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD
ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD WHAT IS A NEURAL NETWORK? The simplest definition of a neural network, more properly referred to as an 'artificial' neural network (ANN), is provided
More informationDeep Neural Networks (1) Hidden layers; Back-propagation
Deep Neural Networs (1) Hidden layers; Bac-propagation Steve Renals Machine Learning Practical MLP Lecture 3 4 October 2017 / 9 October 2017 MLP Lecture 3 Deep Neural Networs (1) 1 Recap: Softmax single
More informationMINIMUM PRECISION REQUIREMENTS FOR THE SVM-SGD LEARNING ALGORITHM. Charbel Sakr, Ameya Patil, Sai Zhang, Yongjune Kim, and Naresh Shanbhag
MINIMUM PRECISION REQUIREMENTS FOR THE SVM-SGD LEARNING ALGORITHM Charbel Sakr, Amea Patil, Sai Zhang, Yongjune Kim, and Naresh Shanbhag Dept. of Electrical and Computer Engineering, Universit of Illinois
More informationThe approximation of piecewise linear membership functions and Łukasiewicz operators
Fuzz Sets and Sstems 54 25 275 286 www.elsevier.com/locate/fss The approimation of piecewise linear membership functions and Łukasiewicz operators József Dombi, Zsolt Gera Universit of Szeged, Institute
More informationPart 8: Neural Networks
METU Informatics Institute Min720 Pattern Classification ith Bio-Medical Applications Part 8: Neural Netors - INTRODUCTION: BIOLOGICAL VS. ARTIFICIAL Biological Neural Netors A Neuron: - A nerve cell as
More informationSpeaker Recognition Using Artificial Neural Networks: RBFNNs vs. EBFNNs
Speaer Recognition Using Artificial Neural Networs: s vs. s BALASKA Nawel ember of the Sstems & Control Research Group within the LRES Lab., Universit 20 Août 55 of Sida, BP: 26, Sida, 21000, Algeria E-mail
More informationAn Accurate Incremental Principal Component Analysis method with capacity of update and downdate
0 International Conference on Computer Science and Information echnolog (ICCSI 0) IPCSI vol. 5 (0) (0) IACSI Press, Singapore DOI: 0.7763/IPCSI.0.V5. An Accurate Incremental Principal Component Analsis
More informationSupervised (BPL) verses Hybrid (RBF) Learning. By: Shahed Shahir
Supervised (BPL) verses Hybrid (RBF) Learning By: Shahed Shahir 1 Outline I. Introduction II. Supervised Learning III. Hybrid Learning IV. BPL Verses RBF V. Supervised verses Hybrid learning VI. Conclusion
More informationNeural Networks and Ensemble Methods for Classification
Neural Networks and Ensemble Methods for Classification NEURAL NETWORKS 2 Neural Networks A neural network is a set of connected input/output units (neurons) where each connection has a weight associated
More informationAn artificial neural networks (ANNs) model is a functional abstraction of the
CHAPER 3 3. Introduction An artificial neural networs (ANNs) model is a functional abstraction of the biological neural structures of the central nervous system. hey are composed of many simple and highly
More informationWhat Do Neural Networks Do? MLP Lecture 3 Multi-layer networks 1
What Do Neural Networks Do? MLP Lecture 3 Multi-layer networks 1 Multi-layer networks Steve Renals Machine Learning Practical MLP Lecture 3 7 October 2015 MLP Lecture 3 Multi-layer networks 2 What Do Single
More informationDeep Neural Networks (1) Hidden layers; Back-propagation
Deep Neural Networs (1) Hidden layers; Bac-propagation Steve Renals Machine Learning Practical MLP Lecture 3 2 October 2018 http://www.inf.ed.ac.u/teaching/courses/mlp/ MLP Lecture 3 / 2 October 2018 Deep
More informationEnhancing Generalization Capability of SVM Classifiers with Feature Weight Adjustment
Enhancing Generalization Capability of SVM Classifiers ith Feature Weight Adjustment Xizhao Wang and Qiang He College of Mathematics and Computer Science, Hebei University, Baoding 07002, Hebei, China
More informationArtificial Neural Networks Examination, June 2005
Artificial Neural Networks Examination, June 2005 Instructions There are SIXTY questions. (The pass mark is 30 out of 60). For each question, please select a maximum of ONE of the given answers (either
More informationArtificial Neural Networks. Historical description
Artificial Neural Networks Historical description Victor G. Lopez 1 / 23 Artificial Neural Networks (ANN) An artificial neural network is a computational model that attempts to emulate the functions of
More informationRevision: Neural Network
Revision: Neural Network Exercise 1 Tell whether each of the following statements is true or false by checking the appropriate box. Statement True False a) A perceptron is guaranteed to perfectly learn
More informationEstimation of the wrist torque of robot gripper using data fusion and ANN techniques
Estimation of the wrist torque of robot gripper using data fusion and ANN techniques Wu Ting 1 Tang Xue-hua 1 Li Zhu 1 School of Mechanical Engineering Shanghai Dianji Universit Shanghai 0040 China School
More informationJournal of Inequalities in Pure and Applied Mathematics
Journal of Inequalities in Pure and Applied Mathematics http://jipam.vu.edu.au/ Volume 2, Issue 2, Article 15, 21 ON SOME FUNDAMENTAL INTEGRAL INEQUALITIES AND THEIR DISCRETE ANALOGUES B.G. PACHPATTE DEPARTMENT
More informationNeural Networks and Fuzzy Logic Rajendra Dept.of CSE ASCET
Unit-. Definition Neural network is a massively parallel distributed processing system, made of highly inter-connected neural computing elements that have the ability to learn and thereby acquire knowledge
More informationNeural Networks. Fundamentals Framework for distributed processing Network topologies Training of ANN s Notation Perceptron Back Propagation
Neural Networks Fundamentals Framework for distributed processing Network topologies Training of ANN s Notation Perceptron Back Propagation Neural Networks Historical Perspective A first wave of interest
More informationNeuro-Fuzzy Comp. Ch. 4 June 1, 2005
Neuro-Fuzz Comp Ch 4 June, 25 4 Feedforward Multilaer Neural Networks part I Feedforward multilaer neural networks (introduced in sec 7) with supervised error correcting learning are used to approximate
More informationArtificial Neural Networks Examination, March 2004
Artificial Neural Networks Examination, March 2004 Instructions There are SIXTY questions (worth up to 60 marks). The exam mark (maximum 60) will be added to the mark obtained in the laborations (maximum
More informationApplication of Artificial Neural Networks in Evaluation and Identification of Electrical Loss in Transformers According to the Energy Consumption
Application of Artificial Neural Networks in Evaluation and Identification of Electrical Loss in Transformers According to the Energy Consumption ANDRÉ NUNES DE SOUZA, JOSÉ ALFREDO C. ULSON, IVAN NUNES
More informationFeedforward Neural Networks
Chapter 4 Feedforward Neural Networks 4. Motivation Let s start with our logistic regression model from before: P(k d) = softma k =k ( λ(k ) + w d λ(k, w) ). (4.) Recall that this model gives us a lot
More informationNeural Networks biological neuron artificial neuron 1
Neural Networks biological neuron artificial neuron 1 A two-layer neural network Output layer (activation represents classification) Weighted connections Hidden layer ( internal representation ) Input
More informationMultilayer Perceptron = FeedForward Neural Network
Multilayer Perceptron = FeedForward Neural Networ History Definition Classification = feedforward operation Learning = bacpropagation = local optimization in the space of weights Pattern Classification
More informationNeural Networks. Chapter 18, Section 7. TB Artificial Intelligence. Slides from AIMA 1/ 21
Neural Networks Chapter 8, Section 7 TB Artificial Intelligence Slides from AIMA http://aima.cs.berkeley.edu / 2 Outline Brains Neural networks Perceptrons Multilayer perceptrons Applications of neural
More informationShort-term Streamflow Forecasting: ARIMA Vs Neural Networks
Short-term Streamflow Forecasting: ARIMA Vs Neural Networks,JUAN FRAUSTO-SOLIS 1, ESMERALDA PITA 2, JAVIER LAGUNAS 2 1 Tecnológico de Monterre, Campus Cuernavaca Autopista del Sol km 14, Colonia Real del
More informationOptimum Structure of Feed Forward Neural Networks by SOM Clustering of Neuron Activations
Optimum Structure of Feed Forward Neural Networks b SOM Clustering of Neuron Activations Samarasinghe, S. Centre for Advanced Computational Solutions, Natural Resources Engineering Group Lincoln Universit,
More informationPattern Recognition Prof. P. S. Sastry Department of Electronics and Communication Engineering Indian Institute of Science, Bangalore
Pattern Recognition Prof. P. S. Sastry Department of Electronics and Communication Engineering Indian Institute of Science, Bangalore Lecture - 27 Multilayer Feedforward Neural networks with Sigmoidal
More informationIn the Name of God. Lecture 11: Single Layer Perceptrons
1 In the Name of God Lecture 11: Single Layer Perceptrons Perceptron: architecture We consider the architecture: feed-forward NN with one layer It is sufficient to study single layer perceptrons with just
More informationMachine Learning and Data Mining. Multi-layer Perceptrons & Neural Networks: Basics. Prof. Alexander Ihler
+ Machine Learning and Data Mining Multi-layer Perceptrons & Neural Networks: Basics Prof. Alexander Ihler Linear Classifiers (Perceptrons) Linear Classifiers a linear classifier is a mapping which partitions
More informationDeep Feedforward Networks
Deep Feedforward Networks Liu Yang March 30, 2017 Liu Yang Short title March 30, 2017 1 / 24 Overview 1 Background A general introduction Example 2 Gradient based learning Cost functions Output Units 3
More informationNeural networks. Chapter 19, Sections 1 5 1
Neural networks Chapter 19, Sections 1 5 Chapter 19, Sections 1 5 1 Outline Brains Neural networks Perceptrons Multilayer perceptrons Applications of neural networks Chapter 19, Sections 1 5 2 Brains 10
More informationSelection of the Appropriate Lag Structure of Foreign Exchange Rates Forecasting Based on Autocorrelation Coefficient
Selection of the Appropriate Lag Structure of Foreign Exchange Rates Forecasting Based on Autocorrelation Coefficient Wei Huang 1,2, Shouyang Wang 2, Hui Zhang 3,4, and Renbin Xiao 1 1 School of Management,
More informationLecture 16: Introduction to Neural Networks
Lecture 16: Introduction to Neural Networs Instructor: Aditya Bhasara Scribe: Philippe David CS 5966/6966: Theory of Machine Learning March 20 th, 2017 Abstract In this lecture, we consider Bacpropagation,
More informationAPPENDIX D Rotation and the General Second-Degree Equation
APPENDIX D Rotation and the General Second-Degree Equation Rotation of Aes Invariants Under Rotation After rotation of the - and -aes counterclockwise through an angle, the rotated aes are denoted as the
More informationIN neural-network training, the most well-known online
IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 10, NO. 1, JANUARY 1999 161 On the Kalman Filtering Method in Neural-Network Training and Pruning John Sum, Chi-sing Leung, Gilbert H. Young, and Wing-kay Kan
More informationNeural Networks (Part 1) Goals for the lecture
Neural Networks (Part ) Mark Craven and David Page Computer Sciences 760 Spring 208 www.biostat.wisc.edu/~craven/cs760/ Some of the slides in these lectures have been adapted/borrowed from materials developed
More informationA Logarithmic Neural Network Architecture for Unbounded Non-Linear Function Approximation
1 Introduction A Logarithmic Neural Network Architecture for Unbounded Non-Linear Function Approximation J Wesley Hines Nuclear Engineering Department The University of Tennessee Knoxville, Tennessee,
More informationNeural Networks, Computation Graphs. CMSC 470 Marine Carpuat
Neural Networks, Computation Graphs CMSC 470 Marine Carpuat Binary Classification with a Multi-layer Perceptron φ A = 1 φ site = 1 φ located = 1 φ Maizuru = 1 φ, = 2 φ in = 1 φ Kyoto = 1 φ priest = 0 φ
More informationArtificial neural networks
Artificial neural networks Chapter 8, Section 7 Artificial Intelligence, spring 203, Peter Ljunglöf; based on AIMA Slides c Stuart Russel and Peter Norvig, 2004 Chapter 8, Section 7 Outline Brains Neural
More informationCS:4420 Artificial Intelligence
CS:4420 Artificial Intelligence Spring 2018 Neural Networks Cesare Tinelli The University of Iowa Copyright 2004 18, Cesare Tinelli and Stuart Russell a a These notes were originally developed by Stuart
More informationA Novel Rejection Measurement in Handwritten Numeral Recognition Based on Linear Discriminant Analysis
009 0th International Conference on Document Analysis and Recognition A Novel Rejection easurement in Handwritten Numeral Recognition Based on Linear Discriminant Analysis Chun Lei He Louisa Lam Ching
More informationMachine Learning for Large-Scale Data Analysis and Decision Making A. Neural Networks Week #6
Machine Learning for Large-Scale Data Analysis and Decision Making 80-629-17A Neural Networks Week #6 Today Neural Networks A. Modeling B. Fitting C. Deep neural networks Today s material is (adapted)
More informationReading Group on Deep Learning Session 1
Reading Group on Deep Learning Session 1 Stephane Lathuiliere & Pablo Mesejo 2 June 2016 1/31 Contents Introduction to Artificial Neural Networks to understand, and to be able to efficiently use, the popular
More informationCONTINUOUS SPATIAL DATA ANALYSIS
CONTINUOUS SPATIAL DATA ANALSIS 1. Overview of Spatial Stochastic Processes The ke difference between continuous spatial data and point patterns is that there is now assumed to be a meaningful value, s
More informationFault Masking By Probabilistic Voting
OncuBilim Algorithm And Sstems Labs. Vol.9, Art.No:,(9) Fault Masking B Probabilistic Voting B. Bakant ALAGÖZ Abstract: In this stud, we introduced a probabilistic voter, regarding smbol probabilities
More informationWeight Initialization Methods for Multilayer Feedforward. 1
Weight Initialization Methods for Multilayer Feedforward. 1 Mercedes Fernández-Redondo - Carlos Hernández-Espinosa. Universidad Jaume I, Campus de Riu Sec, Edificio TI, Departamento de Informática, 12080
More informationCourse 395: Machine Learning - Lectures
Course 395: Machine Learning - Lectures Lecture 1-2: Concept Learning (M. Pantic) Lecture 3-4: Decision Trees & CBC Intro (M. Pantic & S. Petridis) Lecture 5-6: Evaluating Hypotheses (S. Petridis) Lecture
More informationECE521 Lecture 7/8. Logistic Regression
ECE521 Lecture 7/8 Logistic Regression Outline Logistic regression (Continue) A single neuron Learning neural networks Multi-class classification 2 Logistic regression The output of a logistic regression
More informationCSE 546 Midterm Exam, Fall 2014
CSE 546 Midterm Eam, Fall 2014 1. Personal info: Name: UW NetID: Student ID: 2. There should be 14 numbered pages in this eam (including this cover sheet). 3. You can use an material ou brought: an book,
More informationSingle layer NN. Neuron Model
Single layer NN We consider the simple architecture consisting of just one neuron. Generalization to a single layer with more neurons as illustrated below is easy because: M M The output units are independent
More informationECLT 5810 Classification Neural Networks. Reference: Data Mining: Concepts and Techniques By J. Hand, M. Kamber, and J. Pei, Morgan Kaufmann
ECLT 5810 Classification Neural Networks Reference: Data Mining: Concepts and Techniques By J. Hand, M. Kamber, and J. Pei, Morgan Kaufmann Neural Networks A neural network is a set of connected input/output
More informationCombination of M-Estimators and Neural Network Model to Analyze Inside/Outside Bark Tree Diameters
Combination of M-Estimators and Neural Network Model to Analyze Inside/Outside Bark Tree Diameters Kyriaki Kitikidou, Elias Milios, Lazaros Iliadis, and Minas Kaymakis Democritus University of Thrace,
More informationNeural networks. Chapter 20, Section 5 1
Neural networks Chapter 20, Section 5 Chapter 20, Section 5 Outline Brains Neural networks Perceptrons Multilayer perceptrons Applications of neural networks Chapter 20, Section 5 2 Brains 0 neurons of
More informationNeural Networks Introduction
Neural Networks Introduction H.A Talebi Farzaneh Abdollahi Department of Electrical Engineering Amirkabir University of Technology Winter 2011 H. A. Talebi, Farzaneh Abdollahi Neural Networks 1/22 Biological
More informationStability Analysis of a Geometrically Imperfect Structure using a Random Field Model
Stabilit Analsis of a Geometricall Imperfect Structure using a Random Field Model JAN VALEŠ, ZDENĚK KALA Department of Structural Mechanics Brno Universit of Technolog, Facult of Civil Engineering Veveří
More informationDesign Collocation Neural Network to Solve Singular Perturbed Problems with Initial Conditions
Article International Journal of Modern Engineering Sciences, 204, 3(): 29-38 International Journal of Modern Engineering Sciences Journal homepage:www.modernscientificpress.com/journals/ijmes.aspx ISSN:
More informationArtificial Neural Network : Training
Artificial Neural Networ : Training Debasis Samanta IIT Kharagpur debasis.samanta.iitgp@gmail.com 06.04.2018 Debasis Samanta (IIT Kharagpur) Soft Computing Applications 06.04.2018 1 / 49 Learning of neural
More informationMultilayer Neural Networks. (sometimes called Multilayer Perceptrons or MLPs)
Multilayer Neural Networks (sometimes called Multilayer Perceptrons or MLPs) Linear separability Hyperplane In 2D: w x + w 2 x 2 + w 0 = 0 Feature x 2 = w w 2 x w 0 w 2 Feature 2 A perceptron can separate
More informationOn Node-Fault-Injection Training of an RBF Network
On Node-Fault-Injection Training of an RBF Network John Sum 1, Chi-sing Leung 2, and Kevin Ho 3 1 Institute of E-Commerce, National Chung Hsing University Taichung 402, Taiwan pfsum@nchu.edu.tw 2 Department
More informationCSC242: Intro to AI. Lecture 21
CSC242: Intro to AI Lecture 21 Administrivia Project 4 (homeworks 18 & 19) due Mon Apr 16 11:59PM Posters Apr 24 and 26 You need an idea! You need to present it nicely on 2-wide by 4-high landscape pages
More informationIntroduction to Neural Networks
Introduction to Neural Networks What are (Artificial) Neural Networks? Models of the brain and nervous system Highly parallel Process information much more like the brain than a serial computer Learning
More information1 History of statistical/machine learning. 2 Supervised learning. 3 Two approaches to supervised learning. 4 The general learning procedure
Overview Breiman L (2001). Statistical modeling: The two cultures Statistical learning reading group Aleander L 1 Histor of statistical/machine learning 2 Supervised learning 3 Two approaches to supervised
More informationArtificial Neural Networks. Edward Gatt
Artificial Neural Networks Edward Gatt What are Neural Networks? Models of the brain and nervous system Highly parallel Process information much more like the brain than a serial computer Learning Very
More informationMultilayer Neural Networks. (sometimes called Multilayer Perceptrons or MLPs)
Multilayer Neural Networks (sometimes called Multilayer Perceptrons or MLPs) Linear separability Hyperplane In 2D: w 1 x 1 + w 2 x 2 + w 0 = 0 Feature 1 x 2 = w 1 w 2 x 1 w 0 w 2 Feature 2 A perceptron
More informationChapter 6: Backpropagation Nets
Chapter 6: Bacpropagation Nets Architecture: at least one laer of non-linear hidden units Learning: supervised, error driven, generalized delta rule Derivation of the weight update formula (with gradient
More informationArtificial Neural Networks" and Nonparametric Methods" CMPSCI 383 Nov 17, 2011!
Artificial Neural Networks" and Nonparametric Methods" CMPSCI 383 Nov 17, 2011! 1 Todayʼs lecture" How the brain works (!)! Artificial neural networks! Perceptrons! Multilayer feed-forward networks! Error
More informationArtificial Neural Networks
Introduction ANN in Action Final Observations Application: Poverty Detection Artificial Neural Networks Alvaro J. Riascos Villegas University of los Andes and Quantil July 6 2018 Artificial Neural Networks
More informationOn Weight-Noise-Injection Training
On Weight-Noise-Injection Training Kevin Ho 1, Chi-sing Leung 2, and John Sum 3 1 Department of Computer Science and Communication Engineering, Providence University, Sha-Lu, Taiwan. ho@pu.edu.tw 2 Department
More informationNeural networks. Chapter 20. Chapter 20 1
Neural networks Chapter 20 Chapter 20 1 Outline Brains Neural networks Perceptrons Multilayer networks Applications of neural networks Chapter 20 2 Brains 10 11 neurons of > 20 types, 10 14 synapses, 1ms
More informationAN INTRODUCTION TO NEURAL NETWORKS. Scott Kuindersma November 12, 2009
AN INTRODUCTION TO NEURAL NETWORKS Scott Kuindersma November 12, 2009 SUPERVISED LEARNING We are given some training data: We must learn a function If y is discrete, we call it classification If it is
More informationArtificial Intelligence
Artificial Intelligence Jeff Clune Assistant Professor Evolving Artificial Intelligence Laboratory Announcements Be making progress on your projects! Three Types of Learning Unsupervised Supervised Reinforcement
More informationMultilayer Neural Networks
Multilayer Neural Networks Multilayer Neural Networks Discriminant function flexibility NON-Linear But with sets of linear parameters at each layer Provably general function approximators for sufficient
More informationRecursive Least Squares for an Entropy Regularized MSE Cost Function
Recursive Least Squares for an Entropy Regularized MSE Cost Function Deniz Erdogmus, Yadunandana N. Rao, Jose C. Principe Oscar Fontenla-Romero, Amparo Alonso-Betanzos Electrical Eng. Dept., University
More informationModified Learning for Discrete Multi-Valued Neuron
Proceedings of International Joint Conference on Neural Networks, Dallas, Texas, USA, August 4-9, 2013 Modified Learning for Discrete Multi-Valued Neuron Jin-Ping Chen, Shin-Fu Wu, and Shie-Jue Lee Department
More informationMachine Learning (CSE 446): Neural Networks
Machine Learning (CSE 446): Neural Networks Noah Smith c 2017 University of Washington nasmith@cs.washington.edu November 6, 2017 1 / 22 Admin No Wednesday office hours for Noah; no lecture Friday. 2 /
More informationVISION TRACKING PREDICTION
VISION RACKING PREDICION Eri Cuevas,2, Daniel Zaldivar,2, and Raul Rojas Institut für Informati, Freie Universität Berlin, ausstr 9, D-495 Berlin, German el 0049-30-83852485 2 División de Electrónica Computación,
More informationSupervised Learning. George Konidaris
Supervised Learning George Konidaris gdk@cs.brown.edu Fall 2017 Machine Learning Subfield of AI concerned with learning from data. Broadly, using: Experience To Improve Performance On Some Task (Tom Mitchell,
More informationA Novel Activity Detection Method
A Novel Activity Detection Method Gismy George P.G. Student, Department of ECE, Ilahia College of,muvattupuzha, Kerala, India ABSTRACT: This paper presents an approach for activity state recognition of
More informationNonlinear Classification
Nonlinear Classification INFO-4604, Applied Machine Learning University of Colorado Boulder October 5-10, 2017 Prof. Michael Paul Linear Classification Most classifiers we ve seen use linear functions
More informationGraphic Representation Method and Neural Network Recognition of Time-Frequency Vectors of Speech Information
Programming and Computer Software, Vol. 29, No. 4, 2003, pp. 210 218. Translated from Programmirovanie, Vol. 29, No. 4, 2003. Original Russian Text Copright 2003 b Zhirkov, Kortchagine, Lukin, Krlov, Baakovskii.
More informationLast update: October 26, Neural networks. CMSC 421: Section Dana Nau
Last update: October 26, 207 Neural networks CMSC 42: Section 8.7 Dana Nau Outline Applications of neural networks Brains Neural network units Perceptrons Multilayer perceptrons 2 Example Applications
More informationChapter 2 Single Layer Feedforward Networks
Chapter 2 Single Layer Feedforward Networks By Rosenblatt (1962) Perceptrons For modeling visual perception (retina) A feedforward network of three layers of units: Sensory, Association, and Response Learning
More informationEVALUATION OF COMPLEX PERMITTIVITIES OF MULTILAYER DIELECTRIC SUBSTRATES AT MICROWAVE FREQUENCIES USING WAVEGUIDE MEASUREMENTS
EVALUATION OF COMPLEX PERMITTIVITIES OF MULTILAYER DIELECTRIC SUBSTRATES AT MICROWAVE FREQUENCIES USING WAVEGUIDE MEASUREMENTS R. L. Crave, M. D. Deshpande, C. J. Redd 3, and P. I. Tiemsin () NASA Langle
More informationIntroduction: The Perceptron
Introduction: The Perceptron Haim Sompolinsy, MIT October 4, 203 Perceptron Architecture The simplest type of perceptron has a single layer of weights connecting the inputs and output. Formally, the perceptron
More informationIntroduction to Differential Equations. National Chiao Tung University Chun-Jen Tsai 9/14/2011
Introduction to Differential Equations National Chiao Tung Universit Chun-Jen Tsai 9/14/011 Differential Equations Definition: An equation containing the derivatives of one or more dependent variables,
More informationArtificial Neural Network
Artificial Neural Network Contents 2 What is ANN? Biological Neuron Structure of Neuron Types of Neuron Models of Neuron Analogy with human NN Perceptron OCR Multilayer Neural Network Back propagation
More informationVibration of Plate on Foundation with Four Edges Free by Finite Cosine Integral Transform Method
854 Vibration of Plate on Foundation with Four Edges Free b Finite Cosine Integral Transform Method Abstract The analtical solutions for the natural frequencies and mode shapes of the rectangular plate
More informationLecture 4: Feed Forward Neural Networks
Lecture 4: Feed Forward Neural Networks Dr. Roman V Belavkin Middlesex University BIS4435 Biological neurons and the brain A Model of A Single Neuron Neurons as data-driven models Neural Networks Training
More informationObject Recognition Using a Neural Network and Invariant Zernike Features
Object Recognition Using a Neural Network and Invariant Zernike Features Abstract : In this paper, a neural network (NN) based approach for translation, scale, and rotation invariant recognition of objects
More informationPattern Classification
Pattern Classification All materials in these slides were taen from Pattern Classification (2nd ed) by R. O. Duda,, P. E. Hart and D. G. Stor, John Wiley & Sons, 2000 with the permission of the authors
More informationThe perceptron learning algorithm is one of the first procedures proposed for learning in neural network models and is mostly credited to Rosenblatt.
1 The perceptron learning algorithm is one of the first procedures proposed for learning in neural network models and is mostly credited to Rosenblatt. The algorithm applies only to single layer models
More information