Fundamentals of Computational Neuroscience 2e

Similar documents
Multilayer Perceptrons and Backpropagation. Perceptrons. Recap: Perceptrons. Informatics 1 CG: Lecture 6. Mirella Lapata

Multi-layer neural networks

Introduction to the Introduction to Artificial Neural Network

Multilayer neural networks

Neural Networks. Perceptrons and Backpropagation. Silke Bussen-Heyen. 5th of Novemeber Universität Bremen Fachbereich 3. Neural Networks 1 / 17

Admin NEURAL NETWORKS. Perceptron learning algorithm. Our Nervous System 10/25/16. Assignment 7. Class 11/22. Schedule for the rest of the semester

Multilayer Perceptron (MLP)

EEE 241: Linear Systems

Gradient Descent Learning and Backpropagation

Evaluation of classifiers MLPs

Linear Classification, SVMs and Nearest Neighbors

Support Vector Machines. Vibhav Gogate The University of Texas at dallas

CHALMERS, GÖTEBORGS UNIVERSITET. SOLUTIONS to RE-EXAM for ARTIFICIAL NEURAL NETWORKS. COURSE CODES: FFR 135, FIM 720 GU, PhD

Generalized Linear Methods

1 Convex Optimization

18-660: Numerical Methods for Engineering Design and Optimization

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results.

MATH 567: Mathematical Techniques in Data Science Lab 8

Supporting Information

Intro to Visual Recognition

Support Vector Machines

Multigradient for Neural Networks for Equalizers 1

PERFORMANCE COMPARISON BETWEEN BACK PROPAGATION, RPE AND MRPE ALGORITHMS FOR TRAINING MLP NETWORKS

Neural Networks & Learning

Lecture Notes on Linear Regression

Radial-Basis Function Networks

Fundamentals of Neural Networks

Discriminative classifier: Logistic Regression. CS534-Machine Learning

Support Vector Machines

INF 5860 Machine learning for image classification. Lecture 3 : Image classification and regression part II Anne Solberg January 31, 2018

Week 5: Neural Networks

Kernels in Support Vector Machines. Based on lectures of Martin Law, University of Michigan

1 Input-Output Mappings. 2 Hebbian Failure. 3 Delta Rule Success.

Natural Language Processing and Information Retrieval

Relevance Vector Machines Explained

Neural Networks. Class 22: MLSP, Fall 2016 Instructor: Bhiksha Raj

Lecture 23: Artificial neural networks

CS 3710: Visual Recognition Classification and Detection. Adriana Kovashka Department of Computer Science January 13, 2015

10-701/ Machine Learning, Fall 2005 Homework 3

CS246: Mining Massive Datasets Jure Leskovec, Stanford University

Image classification. Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing i them?

Discriminative classifier: Logistic Regression. CS534-Machine Learning

Nonlinear Classifiers II

Neural networks. Nuno Vasconcelos ECE Department, UCSD

Ensemble Methods: Boosting

17 Support Vector Machines

A neural network with localized receptive fields for visual pattern classification

Support Vector Machines

MLE and Bayesian Estimation. Jie Tang Department of Computer Science & Technology Tsinghua University 2012

Lecture 3: Dual problems and Kernels

Non-linear Canonical Correlation Analysis Using a RBF Network

The Study of Teaching-learning-based Optimization Algorithm

Logistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI

CSE 546 Midterm Exam, Fall 2014(with Solution)

Logistic Classifier CISC 5800 Professor Daniel Leeds

Lecture 10 Support Vector Machines. Oct

Neural Networks. Neural Network Motivation. Why Neural Networks? Comments on Blue Gene. More Comments on Blue Gene

Supervised Learning NNs

Online Classification: Perceptron and Winnow

C4B Machine Learning Answers II. = σ(z) (1 σ(z)) 1 1 e z. e z = σ(1 σ) (1 + e z )

Using deep belief network modelling to characterize differences in brain morphometry in schizophrenia

Learning with Tensor Representation

Linear Regression Introduction to Machine Learning. Matt Gormley Lecture 5 September 14, Readings: Bishop, 3.1

CS294A Lecture notes. Andrew Ng

Internet Engineering. Jacek Mazurkiewicz, PhD Softcomputing. Part 3: Recurrent Artificial Neural Networks Self-Organising Artificial Neural Networks

2 Laminar Structure of Cortex. 4 Area Structure of Cortex

The Cortex. Networks. Laminar Structure of Cortex. Chapter 3, O Reilly & Munakata.

Maximal Margin Classifier

Deep Learning. Boyang Albert Li, Jie Jay Tan

Structured Perceptrons & Structural SVMs

arxiv: v1 [cs.cv] 9 Nov 2017

Which Separator? Spring 1

Why feed-forward networks are in a bad shape

CS294A Lecture notes. Andrew Ng

Pattern Classification

Sparse Gaussian Processes Using Backward Elimination

2 STATISTICALLY OPTIMAL TRAINING DATA 2.1 A CRITERION OF OPTIMALITY We revew the crteron of statstcally optmal tranng data (Fukumzu et al., 1994). We

CSC321 Tutorial 9: Review of Boltzmann machines and simulated annealing

Support Vector Machines

Advanced Introduction to Machine Learning

Support Vector Machines. Jie Tang Knowledge Engineering Group Department of Computer Science and Technology Tsinghua University 2012

Multi layer feed-forward NN FFNN. XOR problem. XOR problem. Neural Network for Speech. NETtalk (Sejnowski & Rosenberg, 1987) NETtalk (contd.

Support Vector Machines

We present the algorithm first, then derive it later. Assume access to a dataset {(x i, y i )} n i=1, where x i R d and y i { 1, 1}.

FORECASTING EXCHANGE RATE USING SUPPORT VECTOR MACHINES

CSC 411 / CSC D11 / CSC C11

SDMML HT MSc Problem Sheet 4

Statistical Foundations of Pattern Recognition

UVA CS / Introduc8on to Machine Learning and Data Mining. Lecture 10: Classifica8on with Support Vector Machine (cont.

CoSMo 2012 Gunnar Blohm

Back-Propagation Algorithm. Perceptron Gradient Descent Multilayered neural network Back-Propagation More on Back-Propagation Examples

Other NN Models. Reinforcement learning (RL) Probabilistic neural networks

CHAPTER 3 ARTIFICIAL NEURAL NETWORKS AND LEARNING ALGORITHM

Regularized Discriminant Analysis for Face Recognition

Gaussian process classification: a message-passing viewpoint

CSCI B609: Foundations of Data Science

Transient Stability Assessment of Power System Based on Support Vector Machine

Dynamic Ensemble Selection and Instantaneous Pruning for Regression

The exam is closed book, closed notes except your one-page cheat sheet.

A Fuzzy Image Segmentation using Feedforward Neural Networks with Supervised Learning

Transcription:

Fundamentals of Computatonal Neuroscence e Thomas Trappenberg February 7, 9 Chapter 6: Feed-forward mappng networks

Dgtal representaton of letter A 3 3 4 5 3 33 4 5 34 35 <-5 Optcal character recognton: Predct meanng from features Eg, gven features x, what s the character y f : x S n y S m

Further examples gven by lookup table A Boolean AND functon x x y B Non-Boolean functon x x y 3 - - - - 5 7

The populaton node as perceptron Update rule: r out = g(wr n ) (component-wse: r out = g( P j w jr n j )) For example: r n = x, ỹ = r out, lnear gran functon g(x) = x: ỹ = w x + w x 5 y y, ~ r n r n w w Σ g r out -5 4 x 4 4 x 4

How to fnd the rght weght values? Objectve (error) functon, for example: mean square error (MSE) E = (r out y ) Gradent descent method: w j w j ɛ E w j = w j ɛ(y r out )rj n for MSE, lnear gan E(w) Intalze weghts arbtrarly Repeat untl error s suffcently small Apply a sample pattern to the nput nodes: r = r n = ξ n Calculate rate of the output nodes: r out = g( P j w j rn ) j Compute the delta term for the output layer: δ = g (h out w )(ξ out Update the weght matrx by addng the term: w j = ɛδ r j n r out )

Example: OCR A Tranng pattern B Learnng curve C Generalzaton ablty >> dsplayletter() +++ +++ +++++ ++ ++ ++ ++ +++ +++ +++++++++ +++++++++++ +++ +++ +++ +++ +++ +++ +++ +++ Average Hammng dstance 9 8 7 6 5 4 3 5 5 Tranng step Average Hammng dstance 8 6 4 4 6 8 Number of flpped bts

Example: Boolean functon A Boolean OR functon x x y x x x x w = w = x = w = Θ = Σ y w x + w x = Θ B Boolean XOR functon x x x y x? x x x x

perceptrontranm %% Letter recognton wth threshold perceptron clear; clf; 3 nin=*3; nout=6; 4 wout=rand(nout,nin)-5; 5 6 % tranng vectors 7 load pattern; 8 rin=reshape(pattern, nin, 6); 9 rdes=dag(ones(,6)); % Updatng and tranng network for tranng_step=:; 3 % test all pattern 4 rout=(wout*rin)>5; 5 dsth=sum(sum((rdes-rout)ˆ))/6; 6 error(tranng_step)=dsth; 7 % tranng wth delta rule 8 wout=wout+*(rdes-rout)*rin ; 9 end plot(:9,error) xlabel( Tranng step ) 3 ylabel( Average Hammng dstance )

Multlayer Perceptron (MLP) n h out n n n r h n r n r n rn n out r out rn out w h w out Update rule: r out = g out (w out g h (w h r n )) Learnng rule (error backpropagaton): w j w j ɛ E w j Intalze weghts arbtrarly Repeat untl error s suffcently small Apply a sample pattern to the nput nodes: r := r n = ξ n Propagate nput through the network by calculatng the rates of nodes n successve layers l: r l = g(h l ) = g(p j wl j rl ) j Compute the delta term for the output layer: δ out = g (h out )(ξ out r out ) Back-propagate delta terms through the network: δ l = g (h l ) P j wl j δl j Update weght matrx by addng the term: w j l = ɛδ l rl j

perceptrontranm %% MLP wth backpropagaton learnng on XOR problem clear; clf; 3 N_=; N_h=; N_o=; 4 w_h=rand(n_h,n_)-5; w_o=rand(n_o,n_h)-5; 5 6 % tranng vectors (XOR) 7 r_=[ ; ]; 8 r_d=[ ]; 9 % Updatng and tranng network wth sgmod actvaton functon for sweep=:; % tranng randomly on one pattern 3 =cel(4*rand); 4 r_h=/(+exp(-w_h*r_(:,))); 5 r_o=/(+exp(-w_o*r_h)); 6 d_o=(r_o*(-r_o))*(r_d(:,)-r_o); 7 d_h=(r_h*(-r_h))*(w_o *d_o); 8 w_o=w_o+7*(r_h*d_o ) ; 9 w_h=w_h+7*(r_(:,)*d_h ) ; % test all pattern r_o_test=/(+exp(-w_o*(/(+exp(-w_h*r_))))); d(sweep)=5*sum((r_o_test-r_d)ˆ); 3 end 4 plot(d)

A MLP for representng the XORfuncton B Approxmaton of sn functons by a smal MLP n r 5 5 out 5 r f (x) n r 5 5 Tranng error C Learnng curve for XOR problem 5 4 3 5 Tranng steps 4 6 8 x

Overfttng and underfttng f (x ) 3 overfttng underfttng true mean 3 x Regularzaton, for example E = y ) γ r (r out w

Support Vector Machnes A Lnear large margne classfer D Non-lnear separaton x x B Lnear not separable case C Lnear separable case φ(x)

Further Readngs Smon Haykn (999), Neural networks: a comprehensve foundaton, MacMllan (nd edton) John Hertz, Anders Krogh, and Rchard G Palmer (99), Introducton to the theory of neural computaton, Addson-Wesley Berndt Müller, Joachm Renhardt, and Mchael Thomas Strckland (995), Neural Networks: An Introducton, Sprnger Chrstopher M Bshop (6), Pattern Recognton and Machne Learnng, Sprnger Laurence F Abbott and Sacha B Nelson (), Synaptc plastcty: tamng the beast, n Nature Neurosc (suppl), 3: 78 83 Chrstopher J C Burges (998), A Tutoral on Support Vector Machnes for Pattern Recognton n Data Mnng and Knowledge Dscovery : 67 Alex J Smola and Bernhard Schölhopf (4), A tutoral on support vector regresson n Statstcs and computng 4: 99- Davd E Rumelhart, James L McClelland, and the PDP research group (986), Parallel Dstrbuted Processng: Exploratons n the Mcrostructure of Cognton, MIT Press Peter McLeod, Km Plunkett, and Edmund T Rolls (998), Introducton to connectonst modellng of cogntve processes, Oxford Unversty Press E Bruce Goldsten (999), Sensaton & percepton, Brooks/Cole Publshng Company (5th edton)