COMPARISON STUDY OF SENSITIVITY DEFINITIONS OF NEURAL NETWORKS

Size: px
Start display at page:

Download "COMPARISON STUDY OF SENSITIVITY DEFINITIONS OF NEURAL NETWORKS"

Transcription

1 COMPARISON SUDY OF SENSIIVIY DEFINIIONS OF NEURAL NEORKS CHUN-GUO LI 1, HAI-FENG LI 1 Machine Learning Center, Facult of Mathematics and Computer Science, Hebei Universit, Baoding 07100, China Department of educational administration, Hebei Universit, Baoding 07100, China licg@hbu.cn, lihf@hbu.cn Abstract: his paper compares the sensitivit definitions of neural networs output to input and weight perturbations. Based on the essence of the sensitivit definitions, the authors classif these sensitivit definitions into 3 categories: Noise-to- Signal Ratio, Geometrical propert of derivative, Angle perturbation in Geometr space. he characteristics of these 3 categories of sensitivit definition are discussed respectivel. It is sensible to classif these sensitivit definitions based on the essence of them for researchers can find other new sensitivit definitions of neural networs. Kewords: Sensitivit definition; categories; neural networs; comparison 1. Introduction Sensitivit analsis is an important issue of neural networs research. here had been large amount of sensitivit definitions of different neural networs emerged in the past decades [1-15]. Sensitivit analsis of neural networ s output to input and weight perturbation is expected to settle some appling problems [16-19]. It can be used to predict the effect of perturbations on the networ s output, quantif the error-tolerance and generalization abilities, guide designing a robust networ, and so on. herefore, there are more and more researchers mae much effort to do the sensitivit analsis of neural networ. hese proposed sensitivit definitions are based on distinguished methodologies and on different neural networs. In one hand, some sensitivit definitions are raised to appl on an ensemble of neural networs, while some on one special neural networ. Man sensitivit definitions are brought forward. For one special neural networ (eg. Multilaer perceptron), It is necessar to find some wa to classif these definitions. Consequentl, people can adopt different definition to settle down different problem to get the best results. here are man was to classif the sensitivit definitions. Some people thought that the sensitivit is defined in two approaches [5]: analtical approach and statistical approach. his classification is based on the measure space of the sensitivit definition. Another wa to classif the sensitivit definitions is based on the neural networs that the definition is applied to. Some definitions can be applied to MLP while some limited on Madalines. Some sensitivit definitions can be used for linear neural networs while some for nonlinear neural networs. hese was to classif the sensitivit definitions did not distinguish the essences of those definitions. In this paper, we studied the sensitivit definitions of the neural networs output to input and weight perturbations. According to the essence of the sensitivit definitions, we classified the sensitivit definitions into 3 categories: Noise-to- Signal Ratio, Geometrical propert of derivative, Angle perturbation in Geometr space. Some people proposed the ideas separatel, and some others defined the sensitivit definitions following and improving the idea. his would be sensible for people to find other new sensitivit definitions of neural networs.. Categories of sensitivit definition of neural networs Since the sensitivit of the neural networs emerged, man ideas are proposed to define the sensitivit of neural networs output to input and weight perturbations. Some of these ideas improved the original ideas, while some of them are applied to practical problems. he ideas were reformulated before application. Subsequentl, man new expression formulae appeared. Gathered these formulae together and detected their essences, we classif these sensitivit definitions into 3 approaches according to the borrowed ideas. e describe them in detail in the following 3 sections..1. Noise-to-Signal Ratio

2 Piché [1] raised the idea of Noise-to-Signal for the selection of weight accuracies for Madalines. his idea of sensitivit definition is based on the ratio of noise and signal. Noise, produced from the environment of the signal, affects the transformation of the signal. he Noise-to-Signal Ratio qualifies how much the noise affects the transformation of signal. Based on this idea, Piché regarded the noise to the signal as the perturbations to inputs and weights of the neural networs. He too into account the mathematical characteristics of random variables and proposed the sensitivit definition of neural networs. He defined the sensitivit definition on an Adaline and extended to Madaline. Yeung et al extended Piché s wor on Madalines to MLP and RBFNN [-8]. Given a structured neural networ, all the parameters of the networ are considered as random variables. If the output of the networ can be expressed b, the sensitivit of the networ to input and weight perturbations could be defined as: where NSR (1) is the variance of the output errors caused b weight or first laer input perturbations, and is the variance of the output. Piché derived the calculating formulae of Eq.(1) of Madaline for the input and weight sensitivit in the condition of the following issues: 1. It is assumed that an ensemble of possible ideal neural networs can be identified. hese ideal networs ma represent a general set of Madalines.. he architecture of all ideal neural networs is assumed to be identical. he networ has L laers with N l neurons full connected to the next laer. here is no bias input in the networ. he output of the networ is nonlinear, and all output operators are odd functions. 3. All the weights associated with each node of the networ are modeled as a vector of random variables. Over the ensemble of networs, the weights of a particular laer all have the same variance. he mean value of each weight of the networ over the ensemble is zero. All in all, the components of vector are considered to be zero-mean independent identicall distributed (iid) random variables. he distribution of these random variables is not particularl important. 4. he inputs of the networ are also modeled as a vector of random variables X. he inputs of the networ are also zero-mean independent identicall distributed (iid) random variables. he distribution of these random variables is not particularl important. (Combined the above points 3 and 4, the components of the input and output vector of each laer are independent identicall distributed (iid) random variables.) 5. he mathematical expectation of noise is assumed to be E ln 0 zero. Yeung et al extended Piché s wor on Madalines to MLP and RBFNN [-8]. Although the calculating formulae are reformulated, the sensitivit definition is based on the noise and signal ratio. he differences between them are the procedure to derive the expressions of the numerator and denominator of Noise-to-Signal Ratio. For example, ing et al [] extended the sensitivit definition from Madaline to RBFNN. he considered the sensitivit definition of neural networs output to input perturbations. he relative sensitivit is defined as the following formulae: D X X X SRD () D X where X is the input vector of the neural networ, X the perturbation of the input, D the variance of, the -th output of the networ. ing et al educed the calculating formulae of Eq.() b the calculating properties of variance for RBFNN. Yeung et al [7] reformulated the sensitivit definition and considered Multilaer perceptron with antismmetric squashing activation function. he introduced both the input and weight perturbation of the neural networ. he sensitivit is defined as the following formula: S RD Dg X D g X X D g X where g is the function expression of the networ. Eqs.(-3) were relative sensitivit as the said. Actuall, the two sensitivit formulae are defined in the same wa. he are the expression formulae of Eq.(1). Although these definitions were defined based on the same idea as Eq.(1), the other authors improved the formulae in different expression. For example, beside relative sensitivit definitions, ing et al [,7] also defined the absolute sensitivit definition without the denominator in Eq.(-3), shown as the following two formulae: S D X X X (4) D D (3) S D g X X g X (5) he above absolute sensitivit definition is defined according to the meaning of the variance as mathematical characteristic of random variable. Variance measures the expected deviation amplitude of the value of the random variable from the expected value of the random variable.

3 .. Geometrical propert of derivative First, let us remind some definition in analtical mathematics. Given a differentiable continuous function f x, its first moment can be derived to be f x with respect to x. f x 0 is the slope of function f x in the point x 0, denoting the change rate around x 0. Correspondingl, f x is the function of slope in the definition field of x. Based on this meaning, Zurada et al [9] proposed that the sensitivit of neural networs output to input perturbation can be defined as the derivative of the neural networs. Zurada et al considered a MFNN with a single hidden laer. he networ is regarded as a nonlinear and f x, where differentiable function,,,,,,, 1 x x1 x x n, and n are the number of output and input neurons respectivel. Since is differentiable at x, Jocobian matrix Eq.(6) is derived firstl. he matrix can be evaluated over the entire training set according to several was such as mean square average, absolute value average, and maximum value. Following this sensitivit definition, similarl definitions are proposed to improve and extend the sensitivit definition. ang et al [8] extended this definition to random measure space and defined the sensitivit of RBFNN s output to input regarding the input vector as random vector x1 x x n J x1 x xn (6) x1 x x n he characteristics of this sensitivit definition is listed as the following issues: 1. he neural networs are required to be differentiable and feedforward. his sensitivit definition is unsuitable for perceptron. Because the active function of a perceptron is non-differentiable.. he sensitivit is defined for the neural networs with continuous valued input vector. If the input vector is random vector, this sensitivit definitions would be modified, just as in [8]. 3. Zurada et al defined the sensitivit of the networs outputs to input perturbations. It can be extend to the sensitivit to weight perturbations if the networ is differentiable with respect to the weight variables. 4. his definition sufficientl maes use of the training set to calculate the sensitivit of each input variable. 5. he sensitivit of the input of the networ is calculated for trained networ. herefore, the training algorithm of the networ affects the absolute sensitivit of the input..3. Angle perturbation in Geometr space Marhelen Stevenson et al proposed this idea to define the sensitivit in the geometr space. he too the input and weight perturbation as errors to input and weight. he defined the sensitivit as the probabilit of decision error due to input and weight error. he sensitivit definition is extended from Adaline to Madaline. For an Adaline with n+1 variable inputs X x0 x1 x x n including a threshold input:,,,,, the inputs tae on binar values of either +1 or 1. Associated with the Adaline are n+1 adjustable analog weights: w0, w1, w,, w n. he input-output map of the Adaline can be summarized as 1, 0 output of Adaline X (7) 1, X 0 A Madaline is a laered networ of Adaline elements. herefore the sensitivit definition of Adaline can be extended to Madaline. Marhelen Stevenson et al firstl defined the sensitivit of Adaline. Assuming binar-valued inputs, there are n possible input patterns for an Adaline with n variable inputs. Each input pattern corresponds to a point in n space, which lies on a hpersphere of radius n centered at the origin. here is a hperplane perpendicular to which passes through the origin, where 0 1 x, x, x,, x : X 0. his hperplane, w n, separates the X for which X 0 from those X for which X 0 (i.e. it separates those input vectors which ield a +1 response from those input vectors which ield a 1 response). divides the hpersphere into two hemi- hperspheres. If a randoml oriented perturbation vector is added to, the perturbed weight vector corresponds to another hperplane p p. p would probabl produce decision errors. A decision error is made when an arbitrar input vector is mapped into opposite output categories b and p. he probabilit of a decision error due to an weight perturbation is defined:

4 P Decision Error / (8) where is the angular between and p. It is the sensitivit definition to weight errors. For practical applications, the probabilit of a decision error is derived in terms of the weight perturbation ratio. his definition is based on the Hoff hpersphere-area approximation. he Hoff hpersphere-area approximation states that as n gets large, the points corresponding to the n-dimensional input patterns are approximatel uniforml distributed over the surface of a hpersphere in n space. Consequentl, the percentage of input patterns, which correspond to points on a selected region of the hpersphere, can be approximated as the ratio of the surface content of the selected region to the surface content of the entire hpersphere. his sensitivit definition is convenienc to calculate the sensitivit to weight perturbations. he calculating formulae do not depend on the structure of the networ. hat the sensitivit utilizes is the geometr distribution of the training patterns. he characteristics of this sense of sensitivit definition is listed as the following points: 1. he input vectors are required to be binar-valued, and the output active function of the networ is required to be hard limited function. he active function is non-differentiable. Reminding the limitation of the sensitivit definition, the sensitivit definition here is an alternative definition when the active function of the networ is non-differentiable.. hen n is small (eg. n=10), the points corresponding to the n-dimensional input patterns are hardl possible uniforml distributed over the surface of a hpersphere in n space. In this condition, the calculated probabilit b Eq.(8) would be hard to believe. 3. his stud focused on the sensitivit of Adaline and Madaline output to weight error. For the sensitivit of Adaline and Madaline output to input error, it leaves as the further research, or taes the other wa to define the sensitivit. And for the other neural networs, we should thin over whether this wa to define the sensitivit is suitable or not. 4. he approximating computation formulae are just for large enough n and small enough. Cheng and Yeung [14] also used a geometrical technique to investigate the sensitivit of Neocognitions b modifing the hpersphere model. Unfortunatel the hpersphere model the used fails for the case of the MLP because its input and weight vectors generall do not span a hpersphere surface or volume. 3. Comparison of sensitivit categories he above 3 categories of the sensitivit definitions are modeled based on distinguished methodologies. here are advantages and disadvantages when these definitions are used in applications. he table below gives a straight sight into the comparison of these 3 categories of the sensitivit definitions. Methodolog Noise-to-Signal Ratio Geometr propert of derivative Angle perturbation able 1. Comparison of sensitivit definitions of different methodologies. Input or weight Ensemble of neural perturbation Requirement of Input vector networ the networ Input eight Madaline Adaline Multilaer Madaline Adaline Input and output vector are iid respectivel. Differentiable Binar valued input and output Random Continuous valued Discrete valued From the table above, we can see clearl that different sensitivit definitions are suitable to deal with different ensembles of networs. In application, mabe of the 3 sensitivit definitions are suitable to define the sensitivit definitions of the networs output to input perturbations or weight perturbations. For example, if we are studing the sensitivit of the Adaline s output to weight perturbations, both Noise-to-Signal ratio and angle perturbation can be used to help to sensitivit analsis. But which of them is better to sensitivit analsis is left to further research. 4. Conclusion his paper compares several sensitivit definitions of

5 neural networs output to input and weight perturbations. he sensitivit definitions are classified into 3 categories according to the essence of the sensitivit definition. hese 3 categories of sensitivit definitions can deal with different ensemble of neural networs. his wa of classif the sensitivit classification can help to detect the new sensitivit definition of neural networs output to input and weight perturbations. here are still man points to focus on. From able 1, we can see that it is suitable to define the sensitivit definition of a feedforward neural networ s output to weight perturbation based on all the 3 categories. So, which is the best for the sensitivit analsis in this situation is left for the further research. It is necessar to carr out plent of experiments to help discover what is the best answer of the above problem. References [1] Stephen. Piché: he Selection of eight Accuracies for Madalines. IEEE ransactions on Neural Networs. Vol.6, No., March 1995, [] ing.y. Ng, Daniel S. Yeung, Q. Ran, Eric C.C. sang: Statistical Output Sensitivit to Input and eight Perturbations of Radial Basis Function Neuron Networs. IEEE Proceedings of the International Conferences on Sstem, Man and Cbernetics, 00, [3] Jin Young Choi, Chong-Ho Choi: Sensitivit Analsis of Multilaer Perceptron with Differentiable Activation Functions. IEEE ransactions on Neural Networs. Vol.3, No.1, Januar 199, [4] ing.y. Ng, D.S. Yeung: Electronics Letters, 15 th Ma 003, Vol.39, No. 10, [5] Xiaoqin Zeng, Daniel S. Yeung: Sensitivit Analsis of Multilaer Perceptron to Input and eight Perturbations. IEEE ransactions on Neural Networs. Vol.1, No.6, November 001, [6] Xiaoqin Zeng, Daniel S. Yeung: A Quantified Sensitivit Measure for Multilaer Perceptron to Input Perturbation. Nural Computation, 15, 003, [7] Daniel S. Yeung, Xuequan Sun: Using Function Approximation to Analze the Sensitivit of MLP with Antismmetric Squashing Activation Function. IEEE ransactions on Neural Networs, Vol. 13, No. 1, Januar 00, [8] Xizhao ang, Chunguo Li: A New Definition of Sensitivit for RBFNN and Its Applications to Feature Reduction. Second International Smposium on Neural Networs, LNCS 3496, 005, pp [9] Jace M. Zurada, Alesander Malinowsi, Shiro Usui. Perturbation method for deleting redundant inputs of perceptron networs. Neurocomputing 14 (1997) [10] Sherif Hashem: Sensitivit Analsis for Feedforward Artificial Neural Networs with Differentiable Activation Functions. Proceedings of the 199 International Joint Conference on Neural Networs, Baltimore, IEEE, New Jerse, Vol. I, pp [11] Masato Koda: Neural Networ Learning Based on Stochastic Sensitivit Analsis. IEEE ransactions on Sstems, Man and Cbernetics-Part B: Cbernetics. Vol. 7, No.1, Februar 1997, [1] Lale Ozilmaz, ula Yildirim: Sensitivit Analsis for Conic Section Function Neural Networs. IJCNN (4) 000,44-5 [13] Marhelen Stevenson, Rodne inter, Bernard idrow: Sensitivit of Feedforward Neural Networs to eight Errors. IEEE ransactions on Neural Networs, Vol. 1, No. 1, March 1990, [14] Anton Y. Cheng, Daniel S. Yeung: Sensitivit Analsis of Neocognitron. IEEE ransactions on Sstems, Man and Cbernetics-Part C: Applications and Reviews, Vol. 9, No., Ma 1999, [15] Cesare Alippi, Vincenzo Piuri, Mariagiovanna Sami: Sensitivit to Errors in Artificial Neural Networs: A Behavioral Approach. IEEE ransactions on Circuits and Sstems-I: Fundamental heor and Applications, Vol. 4, No. 6, June 1995, [16] D. Shi, D.S. Yeung, J. Gao. Sensitivit analsis applied to the construction of radial basis function networs. Neural Networs, 18 (005), [17] ing.y. Ng, Daniel S. Yeung. Input dimensionalit reduction for Radial Basis Neural Networ classification problems using sensitivit analsis. Proceedings of the First International Conference on Machine Learning and Cbernetics, Beijing, 00, [18] ing.y. Ng, Roc K.C. Chang, Daniel S. Yeung: Dimensionalit Reduction for Denial of Service Detection Problems Using RBFNN Output Sensitivit. Proceedings of the Second International Conference on Machine Learning and Cbernetics, Xi an, 003, [19] Nicolaos B. Karaiannis: Reformulated Radial Basis Neural Networs rained b Gradient Decent. IEEE ransactions on Neural Networs, Vol. 10, No. 3, Ma 1999,

Simple Neural Nets For Pattern Classification

Simple Neural Nets For Pattern Classification CHAPTER 2 Simple Neural Nets For Pattern Classification Neural Networks General Discussion One of the simplest tasks that neural nets can be trained to perform is pattern classification. In pattern classification

More information

Multilayer Neural Networks

Multilayer Neural Networks Pattern Recognition Multilaer Neural Networs Lecture 4 Prof. Daniel Yeung School of Computer Science and Engineering South China Universit of Technolog Outline Introduction (6.) Artificial Neural Networ

More information

Neural Networks and the Back-propagation Algorithm

Neural Networks and the Back-propagation Algorithm Neural Networks and the Back-propagation Algorithm Francisco S. Melo In these notes, we provide a brief overview of the main concepts concerning neural networks and the back-propagation algorithm. We closely

More information

Multilayer Perceptron

Multilayer Perceptron Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Single Perceptron 3 Boolean Function Learning 4

More information

Artificial Neural Networks The Introduction

Artificial Neural Networks The Introduction Artificial Neural Networks The Introduction 01001110 01100101 01110101 01110010 01101111 01101110 01101111 01110110 01100001 00100000 01110011 01101011 01110101 01110000 01101001 01101110 01100001 00100000

More information

Unit III. A Survey of Neural Network Model

Unit III. A Survey of Neural Network Model Unit III A Survey of Neural Network Model 1 Single Layer Perceptron Perceptron the first adaptive network architecture was invented by Frank Rosenblatt in 1957. It can be used for the classification of

More information

ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD

ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD WHAT IS A NEURAL NETWORK? The simplest definition of a neural network, more properly referred to as an 'artificial' neural network (ANN), is provided

More information

Deep Neural Networks (1) Hidden layers; Back-propagation

Deep Neural Networks (1) Hidden layers; Back-propagation Deep Neural Networs (1) Hidden layers; Bac-propagation Steve Renals Machine Learning Practical MLP Lecture 3 4 October 2017 / 9 October 2017 MLP Lecture 3 Deep Neural Networs (1) 1 Recap: Softmax single

More information

MINIMUM PRECISION REQUIREMENTS FOR THE SVM-SGD LEARNING ALGORITHM. Charbel Sakr, Ameya Patil, Sai Zhang, Yongjune Kim, and Naresh Shanbhag

MINIMUM PRECISION REQUIREMENTS FOR THE SVM-SGD LEARNING ALGORITHM. Charbel Sakr, Ameya Patil, Sai Zhang, Yongjune Kim, and Naresh Shanbhag MINIMUM PRECISION REQUIREMENTS FOR THE SVM-SGD LEARNING ALGORITHM Charbel Sakr, Amea Patil, Sai Zhang, Yongjune Kim, and Naresh Shanbhag Dept. of Electrical and Computer Engineering, Universit of Illinois

More information

The approximation of piecewise linear membership functions and Łukasiewicz operators

The approximation of piecewise linear membership functions and Łukasiewicz operators Fuzz Sets and Sstems 54 25 275 286 www.elsevier.com/locate/fss The approimation of piecewise linear membership functions and Łukasiewicz operators József Dombi, Zsolt Gera Universit of Szeged, Institute

More information

Part 8: Neural Networks

Part 8: Neural Networks METU Informatics Institute Min720 Pattern Classification ith Bio-Medical Applications Part 8: Neural Netors - INTRODUCTION: BIOLOGICAL VS. ARTIFICIAL Biological Neural Netors A Neuron: - A nerve cell as

More information

Speaker Recognition Using Artificial Neural Networks: RBFNNs vs. EBFNNs

Speaker Recognition Using Artificial Neural Networks: RBFNNs vs. EBFNNs Speaer Recognition Using Artificial Neural Networs: s vs. s BALASKA Nawel ember of the Sstems & Control Research Group within the LRES Lab., Universit 20 Août 55 of Sida, BP: 26, Sida, 21000, Algeria E-mail

More information

An Accurate Incremental Principal Component Analysis method with capacity of update and downdate

An Accurate Incremental Principal Component Analysis method with capacity of update and downdate 0 International Conference on Computer Science and Information echnolog (ICCSI 0) IPCSI vol. 5 (0) (0) IACSI Press, Singapore DOI: 0.7763/IPCSI.0.V5. An Accurate Incremental Principal Component Analsis

More information

Supervised (BPL) verses Hybrid (RBF) Learning. By: Shahed Shahir

Supervised (BPL) verses Hybrid (RBF) Learning. By: Shahed Shahir Supervised (BPL) verses Hybrid (RBF) Learning By: Shahed Shahir 1 Outline I. Introduction II. Supervised Learning III. Hybrid Learning IV. BPL Verses RBF V. Supervised verses Hybrid learning VI. Conclusion

More information

Neural Networks and Ensemble Methods for Classification

Neural Networks and Ensemble Methods for Classification Neural Networks and Ensemble Methods for Classification NEURAL NETWORKS 2 Neural Networks A neural network is a set of connected input/output units (neurons) where each connection has a weight associated

More information

An artificial neural networks (ANNs) model is a functional abstraction of the

An artificial neural networks (ANNs) model is a functional abstraction of the CHAPER 3 3. Introduction An artificial neural networs (ANNs) model is a functional abstraction of the biological neural structures of the central nervous system. hey are composed of many simple and highly

More information

What Do Neural Networks Do? MLP Lecture 3 Multi-layer networks 1

What Do Neural Networks Do? MLP Lecture 3 Multi-layer networks 1 What Do Neural Networks Do? MLP Lecture 3 Multi-layer networks 1 Multi-layer networks Steve Renals Machine Learning Practical MLP Lecture 3 7 October 2015 MLP Lecture 3 Multi-layer networks 2 What Do Single

More information

Deep Neural Networks (1) Hidden layers; Back-propagation

Deep Neural Networks (1) Hidden layers; Back-propagation Deep Neural Networs (1) Hidden layers; Bac-propagation Steve Renals Machine Learning Practical MLP Lecture 3 2 October 2018 http://www.inf.ed.ac.u/teaching/courses/mlp/ MLP Lecture 3 / 2 October 2018 Deep

More information

Enhancing Generalization Capability of SVM Classifiers with Feature Weight Adjustment

Enhancing Generalization Capability of SVM Classifiers with Feature Weight Adjustment Enhancing Generalization Capability of SVM Classifiers ith Feature Weight Adjustment Xizhao Wang and Qiang He College of Mathematics and Computer Science, Hebei University, Baoding 07002, Hebei, China

More information

Artificial Neural Networks Examination, June 2005

Artificial Neural Networks Examination, June 2005 Artificial Neural Networks Examination, June 2005 Instructions There are SIXTY questions. (The pass mark is 30 out of 60). For each question, please select a maximum of ONE of the given answers (either

More information

Artificial Neural Networks. Historical description

Artificial Neural Networks. Historical description Artificial Neural Networks Historical description Victor G. Lopez 1 / 23 Artificial Neural Networks (ANN) An artificial neural network is a computational model that attempts to emulate the functions of

More information

Revision: Neural Network

Revision: Neural Network Revision: Neural Network Exercise 1 Tell whether each of the following statements is true or false by checking the appropriate box. Statement True False a) A perceptron is guaranteed to perfectly learn

More information

Estimation of the wrist torque of robot gripper using data fusion and ANN techniques

Estimation of the wrist torque of robot gripper using data fusion and ANN techniques Estimation of the wrist torque of robot gripper using data fusion and ANN techniques Wu Ting 1 Tang Xue-hua 1 Li Zhu 1 School of Mechanical Engineering Shanghai Dianji Universit Shanghai 0040 China School

More information

Journal of Inequalities in Pure and Applied Mathematics

Journal of Inequalities in Pure and Applied Mathematics Journal of Inequalities in Pure and Applied Mathematics http://jipam.vu.edu.au/ Volume 2, Issue 2, Article 15, 21 ON SOME FUNDAMENTAL INTEGRAL INEQUALITIES AND THEIR DISCRETE ANALOGUES B.G. PACHPATTE DEPARTMENT

More information

Neural Networks and Fuzzy Logic Rajendra Dept.of CSE ASCET

Neural Networks and Fuzzy Logic Rajendra Dept.of CSE ASCET Unit-. Definition Neural network is a massively parallel distributed processing system, made of highly inter-connected neural computing elements that have the ability to learn and thereby acquire knowledge

More information

Neural Networks. Fundamentals Framework for distributed processing Network topologies Training of ANN s Notation Perceptron Back Propagation

Neural Networks. Fundamentals Framework for distributed processing Network topologies Training of ANN s Notation Perceptron Back Propagation Neural Networks Fundamentals Framework for distributed processing Network topologies Training of ANN s Notation Perceptron Back Propagation Neural Networks Historical Perspective A first wave of interest

More information

Neuro-Fuzzy Comp. Ch. 4 June 1, 2005

Neuro-Fuzzy Comp. Ch. 4 June 1, 2005 Neuro-Fuzz Comp Ch 4 June, 25 4 Feedforward Multilaer Neural Networks part I Feedforward multilaer neural networks (introduced in sec 7) with supervised error correcting learning are used to approximate

More information

Artificial Neural Networks Examination, March 2004

Artificial Neural Networks Examination, March 2004 Artificial Neural Networks Examination, March 2004 Instructions There are SIXTY questions (worth up to 60 marks). The exam mark (maximum 60) will be added to the mark obtained in the laborations (maximum

More information

Application of Artificial Neural Networks in Evaluation and Identification of Electrical Loss in Transformers According to the Energy Consumption

Application of Artificial Neural Networks in Evaluation and Identification of Electrical Loss in Transformers According to the Energy Consumption Application of Artificial Neural Networks in Evaluation and Identification of Electrical Loss in Transformers According to the Energy Consumption ANDRÉ NUNES DE SOUZA, JOSÉ ALFREDO C. ULSON, IVAN NUNES

More information

Feedforward Neural Networks

Feedforward Neural Networks Chapter 4 Feedforward Neural Networks 4. Motivation Let s start with our logistic regression model from before: P(k d) = softma k =k ( λ(k ) + w d λ(k, w) ). (4.) Recall that this model gives us a lot

More information

Neural Networks biological neuron artificial neuron 1

Neural Networks biological neuron artificial neuron 1 Neural Networks biological neuron artificial neuron 1 A two-layer neural network Output layer (activation represents classification) Weighted connections Hidden layer ( internal representation ) Input

More information

Multilayer Perceptron = FeedForward Neural Network

Multilayer Perceptron = FeedForward Neural Network Multilayer Perceptron = FeedForward Neural Networ History Definition Classification = feedforward operation Learning = bacpropagation = local optimization in the space of weights Pattern Classification

More information

Neural Networks. Chapter 18, Section 7. TB Artificial Intelligence. Slides from AIMA 1/ 21

Neural Networks. Chapter 18, Section 7. TB Artificial Intelligence. Slides from AIMA   1/ 21 Neural Networks Chapter 8, Section 7 TB Artificial Intelligence Slides from AIMA http://aima.cs.berkeley.edu / 2 Outline Brains Neural networks Perceptrons Multilayer perceptrons Applications of neural

More information

Short-term Streamflow Forecasting: ARIMA Vs Neural Networks

Short-term Streamflow Forecasting: ARIMA Vs Neural Networks Short-term Streamflow Forecasting: ARIMA Vs Neural Networks,JUAN FRAUSTO-SOLIS 1, ESMERALDA PITA 2, JAVIER LAGUNAS 2 1 Tecnológico de Monterre, Campus Cuernavaca Autopista del Sol km 14, Colonia Real del

More information

Optimum Structure of Feed Forward Neural Networks by SOM Clustering of Neuron Activations

Optimum Structure of Feed Forward Neural Networks by SOM Clustering of Neuron Activations Optimum Structure of Feed Forward Neural Networks b SOM Clustering of Neuron Activations Samarasinghe, S. Centre for Advanced Computational Solutions, Natural Resources Engineering Group Lincoln Universit,

More information

Pattern Recognition Prof. P. S. Sastry Department of Electronics and Communication Engineering Indian Institute of Science, Bangalore

Pattern Recognition Prof. P. S. Sastry Department of Electronics and Communication Engineering Indian Institute of Science, Bangalore Pattern Recognition Prof. P. S. Sastry Department of Electronics and Communication Engineering Indian Institute of Science, Bangalore Lecture - 27 Multilayer Feedforward Neural networks with Sigmoidal

More information

In the Name of God. Lecture 11: Single Layer Perceptrons

In the Name of God. Lecture 11: Single Layer Perceptrons 1 In the Name of God Lecture 11: Single Layer Perceptrons Perceptron: architecture We consider the architecture: feed-forward NN with one layer It is sufficient to study single layer perceptrons with just

More information

Machine Learning and Data Mining. Multi-layer Perceptrons & Neural Networks: Basics. Prof. Alexander Ihler

Machine Learning and Data Mining. Multi-layer Perceptrons & Neural Networks: Basics. Prof. Alexander Ihler + Machine Learning and Data Mining Multi-layer Perceptrons & Neural Networks: Basics Prof. Alexander Ihler Linear Classifiers (Perceptrons) Linear Classifiers a linear classifier is a mapping which partitions

More information

Deep Feedforward Networks

Deep Feedforward Networks Deep Feedforward Networks Liu Yang March 30, 2017 Liu Yang Short title March 30, 2017 1 / 24 Overview 1 Background A general introduction Example 2 Gradient based learning Cost functions Output Units 3

More information

Neural networks. Chapter 19, Sections 1 5 1

Neural networks. Chapter 19, Sections 1 5 1 Neural networks Chapter 19, Sections 1 5 Chapter 19, Sections 1 5 1 Outline Brains Neural networks Perceptrons Multilayer perceptrons Applications of neural networks Chapter 19, Sections 1 5 2 Brains 10

More information

Selection of the Appropriate Lag Structure of Foreign Exchange Rates Forecasting Based on Autocorrelation Coefficient

Selection of the Appropriate Lag Structure of Foreign Exchange Rates Forecasting Based on Autocorrelation Coefficient Selection of the Appropriate Lag Structure of Foreign Exchange Rates Forecasting Based on Autocorrelation Coefficient Wei Huang 1,2, Shouyang Wang 2, Hui Zhang 3,4, and Renbin Xiao 1 1 School of Management,

More information

Lecture 16: Introduction to Neural Networks

Lecture 16: Introduction to Neural Networks Lecture 16: Introduction to Neural Networs Instructor: Aditya Bhasara Scribe: Philippe David CS 5966/6966: Theory of Machine Learning March 20 th, 2017 Abstract In this lecture, we consider Bacpropagation,

More information

APPENDIX D Rotation and the General Second-Degree Equation

APPENDIX D Rotation and the General Second-Degree Equation APPENDIX D Rotation and the General Second-Degree Equation Rotation of Aes Invariants Under Rotation After rotation of the - and -aes counterclockwise through an angle, the rotated aes are denoted as the

More information

IN neural-network training, the most well-known online

IN neural-network training, the most well-known online IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 10, NO. 1, JANUARY 1999 161 On the Kalman Filtering Method in Neural-Network Training and Pruning John Sum, Chi-sing Leung, Gilbert H. Young, and Wing-kay Kan

More information

Neural Networks (Part 1) Goals for the lecture

Neural Networks (Part 1) Goals for the lecture Neural Networks (Part ) Mark Craven and David Page Computer Sciences 760 Spring 208 www.biostat.wisc.edu/~craven/cs760/ Some of the slides in these lectures have been adapted/borrowed from materials developed

More information

A Logarithmic Neural Network Architecture for Unbounded Non-Linear Function Approximation

A Logarithmic Neural Network Architecture for Unbounded Non-Linear Function Approximation 1 Introduction A Logarithmic Neural Network Architecture for Unbounded Non-Linear Function Approximation J Wesley Hines Nuclear Engineering Department The University of Tennessee Knoxville, Tennessee,

More information

Neural Networks, Computation Graphs. CMSC 470 Marine Carpuat

Neural Networks, Computation Graphs. CMSC 470 Marine Carpuat Neural Networks, Computation Graphs CMSC 470 Marine Carpuat Binary Classification with a Multi-layer Perceptron φ A = 1 φ site = 1 φ located = 1 φ Maizuru = 1 φ, = 2 φ in = 1 φ Kyoto = 1 φ priest = 0 φ

More information

Artificial neural networks

Artificial neural networks Artificial neural networks Chapter 8, Section 7 Artificial Intelligence, spring 203, Peter Ljunglöf; based on AIMA Slides c Stuart Russel and Peter Norvig, 2004 Chapter 8, Section 7 Outline Brains Neural

More information

CS:4420 Artificial Intelligence

CS:4420 Artificial Intelligence CS:4420 Artificial Intelligence Spring 2018 Neural Networks Cesare Tinelli The University of Iowa Copyright 2004 18, Cesare Tinelli and Stuart Russell a a These notes were originally developed by Stuart

More information

A Novel Rejection Measurement in Handwritten Numeral Recognition Based on Linear Discriminant Analysis

A Novel Rejection Measurement in Handwritten Numeral Recognition Based on Linear Discriminant Analysis 009 0th International Conference on Document Analysis and Recognition A Novel Rejection easurement in Handwritten Numeral Recognition Based on Linear Discriminant Analysis Chun Lei He Louisa Lam Ching

More information

Machine Learning for Large-Scale Data Analysis and Decision Making A. Neural Networks Week #6

Machine Learning for Large-Scale Data Analysis and Decision Making A. Neural Networks Week #6 Machine Learning for Large-Scale Data Analysis and Decision Making 80-629-17A Neural Networks Week #6 Today Neural Networks A. Modeling B. Fitting C. Deep neural networks Today s material is (adapted)

More information

Reading Group on Deep Learning Session 1

Reading Group on Deep Learning Session 1 Reading Group on Deep Learning Session 1 Stephane Lathuiliere & Pablo Mesejo 2 June 2016 1/31 Contents Introduction to Artificial Neural Networks to understand, and to be able to efficiently use, the popular

More information

CONTINUOUS SPATIAL DATA ANALYSIS

CONTINUOUS SPATIAL DATA ANALYSIS CONTINUOUS SPATIAL DATA ANALSIS 1. Overview of Spatial Stochastic Processes The ke difference between continuous spatial data and point patterns is that there is now assumed to be a meaningful value, s

More information

Fault Masking By Probabilistic Voting

Fault Masking By Probabilistic Voting OncuBilim Algorithm And Sstems Labs. Vol.9, Art.No:,(9) Fault Masking B Probabilistic Voting B. Bakant ALAGÖZ Abstract: In this stud, we introduced a probabilistic voter, regarding smbol probabilities

More information

Weight Initialization Methods for Multilayer Feedforward. 1

Weight Initialization Methods for Multilayer Feedforward. 1 Weight Initialization Methods for Multilayer Feedforward. 1 Mercedes Fernández-Redondo - Carlos Hernández-Espinosa. Universidad Jaume I, Campus de Riu Sec, Edificio TI, Departamento de Informática, 12080

More information

Course 395: Machine Learning - Lectures

Course 395: Machine Learning - Lectures Course 395: Machine Learning - Lectures Lecture 1-2: Concept Learning (M. Pantic) Lecture 3-4: Decision Trees & CBC Intro (M. Pantic & S. Petridis) Lecture 5-6: Evaluating Hypotheses (S. Petridis) Lecture

More information

ECE521 Lecture 7/8. Logistic Regression

ECE521 Lecture 7/8. Logistic Regression ECE521 Lecture 7/8 Logistic Regression Outline Logistic regression (Continue) A single neuron Learning neural networks Multi-class classification 2 Logistic regression The output of a logistic regression

More information

CSE 546 Midterm Exam, Fall 2014

CSE 546 Midterm Exam, Fall 2014 CSE 546 Midterm Eam, Fall 2014 1. Personal info: Name: UW NetID: Student ID: 2. There should be 14 numbered pages in this eam (including this cover sheet). 3. You can use an material ou brought: an book,

More information

Single layer NN. Neuron Model

Single layer NN. Neuron Model Single layer NN We consider the simple architecture consisting of just one neuron. Generalization to a single layer with more neurons as illustrated below is easy because: M M The output units are independent

More information

ECLT 5810 Classification Neural Networks. Reference: Data Mining: Concepts and Techniques By J. Hand, M. Kamber, and J. Pei, Morgan Kaufmann

ECLT 5810 Classification Neural Networks. Reference: Data Mining: Concepts and Techniques By J. Hand, M. Kamber, and J. Pei, Morgan Kaufmann ECLT 5810 Classification Neural Networks Reference: Data Mining: Concepts and Techniques By J. Hand, M. Kamber, and J. Pei, Morgan Kaufmann Neural Networks A neural network is a set of connected input/output

More information

Combination of M-Estimators and Neural Network Model to Analyze Inside/Outside Bark Tree Diameters

Combination of M-Estimators and Neural Network Model to Analyze Inside/Outside Bark Tree Diameters Combination of M-Estimators and Neural Network Model to Analyze Inside/Outside Bark Tree Diameters Kyriaki Kitikidou, Elias Milios, Lazaros Iliadis, and Minas Kaymakis Democritus University of Thrace,

More information

Neural networks. Chapter 20, Section 5 1

Neural networks. Chapter 20, Section 5 1 Neural networks Chapter 20, Section 5 Chapter 20, Section 5 Outline Brains Neural networks Perceptrons Multilayer perceptrons Applications of neural networks Chapter 20, Section 5 2 Brains 0 neurons of

More information

Neural Networks Introduction

Neural Networks Introduction Neural Networks Introduction H.A Talebi Farzaneh Abdollahi Department of Electrical Engineering Amirkabir University of Technology Winter 2011 H. A. Talebi, Farzaneh Abdollahi Neural Networks 1/22 Biological

More information

Stability Analysis of a Geometrically Imperfect Structure using a Random Field Model

Stability Analysis of a Geometrically Imperfect Structure using a Random Field Model Stabilit Analsis of a Geometricall Imperfect Structure using a Random Field Model JAN VALEŠ, ZDENĚK KALA Department of Structural Mechanics Brno Universit of Technolog, Facult of Civil Engineering Veveří

More information

Design Collocation Neural Network to Solve Singular Perturbed Problems with Initial Conditions

Design Collocation Neural Network to Solve Singular Perturbed Problems with Initial Conditions Article International Journal of Modern Engineering Sciences, 204, 3(): 29-38 International Journal of Modern Engineering Sciences Journal homepage:www.modernscientificpress.com/journals/ijmes.aspx ISSN:

More information

Artificial Neural Network : Training

Artificial Neural Network : Training Artificial Neural Networ : Training Debasis Samanta IIT Kharagpur debasis.samanta.iitgp@gmail.com 06.04.2018 Debasis Samanta (IIT Kharagpur) Soft Computing Applications 06.04.2018 1 / 49 Learning of neural

More information

Multilayer Neural Networks. (sometimes called Multilayer Perceptrons or MLPs)

Multilayer Neural Networks. (sometimes called Multilayer Perceptrons or MLPs) Multilayer Neural Networks (sometimes called Multilayer Perceptrons or MLPs) Linear separability Hyperplane In 2D: w x + w 2 x 2 + w 0 = 0 Feature x 2 = w w 2 x w 0 w 2 Feature 2 A perceptron can separate

More information

On Node-Fault-Injection Training of an RBF Network

On Node-Fault-Injection Training of an RBF Network On Node-Fault-Injection Training of an RBF Network John Sum 1, Chi-sing Leung 2, and Kevin Ho 3 1 Institute of E-Commerce, National Chung Hsing University Taichung 402, Taiwan pfsum@nchu.edu.tw 2 Department

More information

CSC242: Intro to AI. Lecture 21

CSC242: Intro to AI. Lecture 21 CSC242: Intro to AI Lecture 21 Administrivia Project 4 (homeworks 18 & 19) due Mon Apr 16 11:59PM Posters Apr 24 and 26 You need an idea! You need to present it nicely on 2-wide by 4-high landscape pages

More information

Introduction to Neural Networks

Introduction to Neural Networks Introduction to Neural Networks What are (Artificial) Neural Networks? Models of the brain and nervous system Highly parallel Process information much more like the brain than a serial computer Learning

More information

1 History of statistical/machine learning. 2 Supervised learning. 3 Two approaches to supervised learning. 4 The general learning procedure

1 History of statistical/machine learning. 2 Supervised learning. 3 Two approaches to supervised learning. 4 The general learning procedure Overview Breiman L (2001). Statistical modeling: The two cultures Statistical learning reading group Aleander L 1 Histor of statistical/machine learning 2 Supervised learning 3 Two approaches to supervised

More information

Artificial Neural Networks. Edward Gatt

Artificial Neural Networks. Edward Gatt Artificial Neural Networks Edward Gatt What are Neural Networks? Models of the brain and nervous system Highly parallel Process information much more like the brain than a serial computer Learning Very

More information

Multilayer Neural Networks. (sometimes called Multilayer Perceptrons or MLPs)

Multilayer Neural Networks. (sometimes called Multilayer Perceptrons or MLPs) Multilayer Neural Networks (sometimes called Multilayer Perceptrons or MLPs) Linear separability Hyperplane In 2D: w 1 x 1 + w 2 x 2 + w 0 = 0 Feature 1 x 2 = w 1 w 2 x 1 w 0 w 2 Feature 2 A perceptron

More information

Chapter 6: Backpropagation Nets

Chapter 6: Backpropagation Nets Chapter 6: Bacpropagation Nets Architecture: at least one laer of non-linear hidden units Learning: supervised, error driven, generalized delta rule Derivation of the weight update formula (with gradient

More information

Artificial Neural Networks" and Nonparametric Methods" CMPSCI 383 Nov 17, 2011!

Artificial Neural Networks and Nonparametric Methods CMPSCI 383 Nov 17, 2011! Artificial Neural Networks" and Nonparametric Methods" CMPSCI 383 Nov 17, 2011! 1 Todayʼs lecture" How the brain works (!)! Artificial neural networks! Perceptrons! Multilayer feed-forward networks! Error

More information

Artificial Neural Networks

Artificial Neural Networks Introduction ANN in Action Final Observations Application: Poverty Detection Artificial Neural Networks Alvaro J. Riascos Villegas University of los Andes and Quantil July 6 2018 Artificial Neural Networks

More information

On Weight-Noise-Injection Training

On Weight-Noise-Injection Training On Weight-Noise-Injection Training Kevin Ho 1, Chi-sing Leung 2, and John Sum 3 1 Department of Computer Science and Communication Engineering, Providence University, Sha-Lu, Taiwan. ho@pu.edu.tw 2 Department

More information

Neural networks. Chapter 20. Chapter 20 1

Neural networks. Chapter 20. Chapter 20 1 Neural networks Chapter 20 Chapter 20 1 Outline Brains Neural networks Perceptrons Multilayer networks Applications of neural networks Chapter 20 2 Brains 10 11 neurons of > 20 types, 10 14 synapses, 1ms

More information

AN INTRODUCTION TO NEURAL NETWORKS. Scott Kuindersma November 12, 2009

AN INTRODUCTION TO NEURAL NETWORKS. Scott Kuindersma November 12, 2009 AN INTRODUCTION TO NEURAL NETWORKS Scott Kuindersma November 12, 2009 SUPERVISED LEARNING We are given some training data: We must learn a function If y is discrete, we call it classification If it is

More information

Artificial Intelligence

Artificial Intelligence Artificial Intelligence Jeff Clune Assistant Professor Evolving Artificial Intelligence Laboratory Announcements Be making progress on your projects! Three Types of Learning Unsupervised Supervised Reinforcement

More information

Multilayer Neural Networks

Multilayer Neural Networks Multilayer Neural Networks Multilayer Neural Networks Discriminant function flexibility NON-Linear But with sets of linear parameters at each layer Provably general function approximators for sufficient

More information

Recursive Least Squares for an Entropy Regularized MSE Cost Function

Recursive Least Squares for an Entropy Regularized MSE Cost Function Recursive Least Squares for an Entropy Regularized MSE Cost Function Deniz Erdogmus, Yadunandana N. Rao, Jose C. Principe Oscar Fontenla-Romero, Amparo Alonso-Betanzos Electrical Eng. Dept., University

More information

Modified Learning for Discrete Multi-Valued Neuron

Modified Learning for Discrete Multi-Valued Neuron Proceedings of International Joint Conference on Neural Networks, Dallas, Texas, USA, August 4-9, 2013 Modified Learning for Discrete Multi-Valued Neuron Jin-Ping Chen, Shin-Fu Wu, and Shie-Jue Lee Department

More information

Machine Learning (CSE 446): Neural Networks

Machine Learning (CSE 446): Neural Networks Machine Learning (CSE 446): Neural Networks Noah Smith c 2017 University of Washington nasmith@cs.washington.edu November 6, 2017 1 / 22 Admin No Wednesday office hours for Noah; no lecture Friday. 2 /

More information

VISION TRACKING PREDICTION

VISION TRACKING PREDICTION VISION RACKING PREDICION Eri Cuevas,2, Daniel Zaldivar,2, and Raul Rojas Institut für Informati, Freie Universität Berlin, ausstr 9, D-495 Berlin, German el 0049-30-83852485 2 División de Electrónica Computación,

More information

Supervised Learning. George Konidaris

Supervised Learning. George Konidaris Supervised Learning George Konidaris gdk@cs.brown.edu Fall 2017 Machine Learning Subfield of AI concerned with learning from data. Broadly, using: Experience To Improve Performance On Some Task (Tom Mitchell,

More information

A Novel Activity Detection Method

A Novel Activity Detection Method A Novel Activity Detection Method Gismy George P.G. Student, Department of ECE, Ilahia College of,muvattupuzha, Kerala, India ABSTRACT: This paper presents an approach for activity state recognition of

More information

Nonlinear Classification

Nonlinear Classification Nonlinear Classification INFO-4604, Applied Machine Learning University of Colorado Boulder October 5-10, 2017 Prof. Michael Paul Linear Classification Most classifiers we ve seen use linear functions

More information

Graphic Representation Method and Neural Network Recognition of Time-Frequency Vectors of Speech Information

Graphic Representation Method and Neural Network Recognition of Time-Frequency Vectors of Speech Information Programming and Computer Software, Vol. 29, No. 4, 2003, pp. 210 218. Translated from Programmirovanie, Vol. 29, No. 4, 2003. Original Russian Text Copright 2003 b Zhirkov, Kortchagine, Lukin, Krlov, Baakovskii.

More information

Last update: October 26, Neural networks. CMSC 421: Section Dana Nau

Last update: October 26, Neural networks. CMSC 421: Section Dana Nau Last update: October 26, 207 Neural networks CMSC 42: Section 8.7 Dana Nau Outline Applications of neural networks Brains Neural network units Perceptrons Multilayer perceptrons 2 Example Applications

More information

Chapter 2 Single Layer Feedforward Networks

Chapter 2 Single Layer Feedforward Networks Chapter 2 Single Layer Feedforward Networks By Rosenblatt (1962) Perceptrons For modeling visual perception (retina) A feedforward network of three layers of units: Sensory, Association, and Response Learning

More information

EVALUATION OF COMPLEX PERMITTIVITIES OF MULTILAYER DIELECTRIC SUBSTRATES AT MICROWAVE FREQUENCIES USING WAVEGUIDE MEASUREMENTS

EVALUATION OF COMPLEX PERMITTIVITIES OF MULTILAYER DIELECTRIC SUBSTRATES AT MICROWAVE FREQUENCIES USING WAVEGUIDE MEASUREMENTS EVALUATION OF COMPLEX PERMITTIVITIES OF MULTILAYER DIELECTRIC SUBSTRATES AT MICROWAVE FREQUENCIES USING WAVEGUIDE MEASUREMENTS R. L. Crave, M. D. Deshpande, C. J. Redd 3, and P. I. Tiemsin () NASA Langle

More information

Introduction: The Perceptron

Introduction: The Perceptron Introduction: The Perceptron Haim Sompolinsy, MIT October 4, 203 Perceptron Architecture The simplest type of perceptron has a single layer of weights connecting the inputs and output. Formally, the perceptron

More information

Introduction to Differential Equations. National Chiao Tung University Chun-Jen Tsai 9/14/2011

Introduction to Differential Equations. National Chiao Tung University Chun-Jen Tsai 9/14/2011 Introduction to Differential Equations National Chiao Tung Universit Chun-Jen Tsai 9/14/011 Differential Equations Definition: An equation containing the derivatives of one or more dependent variables,

More information

Artificial Neural Network

Artificial Neural Network Artificial Neural Network Contents 2 What is ANN? Biological Neuron Structure of Neuron Types of Neuron Models of Neuron Analogy with human NN Perceptron OCR Multilayer Neural Network Back propagation

More information

Vibration of Plate on Foundation with Four Edges Free by Finite Cosine Integral Transform Method

Vibration of Plate on Foundation with Four Edges Free by Finite Cosine Integral Transform Method 854 Vibration of Plate on Foundation with Four Edges Free b Finite Cosine Integral Transform Method Abstract The analtical solutions for the natural frequencies and mode shapes of the rectangular plate

More information

Lecture 4: Feed Forward Neural Networks

Lecture 4: Feed Forward Neural Networks Lecture 4: Feed Forward Neural Networks Dr. Roman V Belavkin Middlesex University BIS4435 Biological neurons and the brain A Model of A Single Neuron Neurons as data-driven models Neural Networks Training

More information

Object Recognition Using a Neural Network and Invariant Zernike Features

Object Recognition Using a Neural Network and Invariant Zernike Features Object Recognition Using a Neural Network and Invariant Zernike Features Abstract : In this paper, a neural network (NN) based approach for translation, scale, and rotation invariant recognition of objects

More information

Pattern Classification

Pattern Classification Pattern Classification All materials in these slides were taen from Pattern Classification (2nd ed) by R. O. Duda,, P. E. Hart and D. G. Stor, John Wiley & Sons, 2000 with the permission of the authors

More information

The perceptron learning algorithm is one of the first procedures proposed for learning in neural network models and is mostly credited to Rosenblatt.

The perceptron learning algorithm is one of the first procedures proposed for learning in neural network models and is mostly credited to Rosenblatt. 1 The perceptron learning algorithm is one of the first procedures proposed for learning in neural network models and is mostly credited to Rosenblatt. The algorithm applies only to single layer models

More information