Νεςπο-Ασαυήρ Υπολογιστική Neuro-Fuzzy Computing

Similar documents
Νεςπο-Ασαυήρ Υπολογιστική Neuro-Fuzzy Computing

Simple neuron model Components of simple neuron

Revision: Neural Network

EEE 241: Linear Systems

Simple Neural Nets For Pattern Classification

Single layer NN. Neuron Model

Artifical Neural Networks

CSC Neural Networks. Perceptron Learning Rule

Pattern Recognition Prof. P. S. Sastry Department of Electronics and Communication Engineering Indian Institute of Science, Bangalore

PMR5406 Redes Neurais e Lógica Fuzzy Aula 3 Single Layer Percetron

Artificial Neural Networks. Historical description

The perceptron learning algorithm is one of the first procedures proposed for learning in neural network models and is mostly credited to Rosenblatt.

Neural Networks and the Back-propagation Algorithm

Security Analytics. Topic 6: Perceptron and Support Vector Machine

In the Name of God. Lecture 11: Single Layer Perceptrons

Lecture 4: Perceptrons and Multilayer Perceptrons

ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD

COGS Q250 Fall Homework 7: Learning in Neural Networks Due: 9:00am, Friday 2nd November.

COMP9444 Neural Networks and Deep Learning 2. Perceptrons. COMP9444 c Alan Blair, 2017

1. A discrete-time recurrent network is described by the following equation: y(n + 1) = A y(n) + B x(n)

Neural Networks Lecture 2:Single Layer Classifiers

Data Mining Part 5. Prediction

COMP-4360 Machine Learning Neural Networks

Neural Networks. CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington

Neural Networks (Part 1) Goals for the lecture

Neural networks. Chapter 19, Sections 1 5 1

ADALINE for Pattern Classification

Feedforward Neural Nets and Backpropagation

Artificial Neural Networks Examination, June 2005

Chapter ML:VI. VI. Neural Networks. Perceptron Learning Gradient Descent Multilayer Perceptron Radial Basis Functions

CSC321 Lecture 4 The Perceptron Algorithm

Lecture 7 Artificial neural networks: Supervised learning

Artificial Neural Network

Chapter 2 Single Layer Feedforward Networks

Artificial Neural Networks" and Nonparametric Methods" CMPSCI 383 Nov 17, 2011!

Artificial Neural Networks Examination, June 2004

Engineering Part IIB: Module 4F10 Statistical Pattern Processing Lecture 5: Single Layer Perceptrons & Estimating Linear Classifiers

Multilayer Perceptron

Introduction to Natural Computation. Lecture 9. Multilayer Perceptrons and Backpropagation. Peter Lewis

Lecture 6. Regression

Neural networks. Chapter 20. Chapter 20 1

Multilayer Perceptrons and Backpropagation

NONLINEAR CLASSIFICATION AND REGRESSION. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

SPSS, University of Texas at Arlington. Topics in Machine Learning-EE 5359 Neural Networks

) (d o f. For the previous layer in a neural network (just the rightmost layer if a single neuron), the required update equation is: 2.

Machine Learning

Neural Networks. Nicholas Ruozzi University of Texas at Dallas

Computational Intelligence Lecture 3: Simple Neural Networks for Pattern Classification

Vote. Vote on timing for night section: Option 1 (what we have now) Option 2. Lecture, 6:10-7:50 25 minute dinner break Tutorial, 8:15-9

Artificial neural networks

Multi-layer Perceptron Networks

Lecture 4: Feed Forward Neural Networks

Course 395: Machine Learning - Lectures

CSE 352 (AI) LECTURE NOTES Professor Anita Wasilewska. NEURAL NETWORKS Learning

Numerical Learning Algorithms

Introduction to Artificial Neural Networks

Advanced statistical methods for data analysis Lecture 2

Artificial Neural Networks. Part 2

Classification with Perceptrons. Reading:

Perceptron. (c) Marcin Sydow. Summary. Perceptron

Simulating Neural Networks. Lawrence Ward P465A

Introduction To Artificial Neural Networks

COMP 551 Applied Machine Learning Lecture 14: Neural Networks

Neural Networks and Fuzzy Logic Rajendra Dept.of CSE ASCET

Unit III. A Survey of Neural Network Model

AN INTRODUCTION TO NEURAL NETWORKS. Scott Kuindersma November 12, 2009

Artificial Neural Networks Examination, March 2004

VC-dimension of a context-dependent perceptron

Learning Deep Architectures for AI. Part I - Vijay Chakilam

Artificial Neural Networks. Q550: Models in Cognitive Science Lecture 5

Machine Learning. Neural Networks. Le Song. CSE6740/CS7641/ISYE6740, Fall Lecture 7, September 11, 2012 Based on slides from Eric Xing, CMU

2015 Todd Neller. A.I.M.A. text figures 1995 Prentice Hall. Used by permission. Neural Networks. Todd W. Neller

Linear discriminant functions

Neural networks. Chapter 20, Section 5 1

Multilayer Neural Networks

Computational Intelligence Winter Term 2009/10

Neural Networks. Chapter 18, Section 7. TB Artificial Intelligence. Slides from AIMA 1/ 21

ECE521 Lectures 9 Fully Connected Neural Networks

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Artificial Neural Networks

Artificial Neural Networks The Introduction

Lab 5: 16 th April Exercises on Neural Networks

LECTURE NOTE #NEW 6 PROF. ALAN YUILLE

Neural Networks DWML, /25

ARTIFICIAL INTELLIGENCE. Artificial Neural Networks

Computational Intelligence

1 What a Neural Network Computes

y(x n, w) t n 2. (1)

Computational Intelligence Winter Term 2017/18

1 Machine Learning Concepts (16 points)

CSC321 Lecture 5 Learning in a Single Neuron

Learning and Memory in Neural Networks

100 inference steps doesn't seem like enough. Many neuron-like threshold switching units. Many weighted interconnections among units

Artificial Intelligence

Multi-layer Neural Networks

Machine Learning and Data Mining. Multi-layer Perceptrons & Neural Networks: Basics. Prof. Alexander Ihler

Gradient Descent Training Rule: The Details

The Perceptron Algorithm 1

Backpropagation Neural Net

Neural Networks Lecture 4: Radial Bases Function Networks

Transcription:

Νεςπο-Ασαυήρ Υπολογιστική Neuro-Fuzzy Computing ΗΥ418 Διδάσκων Δημήτριος Κατσαρός @ Τμ. ΗΜΜΥ Πανεπιστήμιο Θεσσαλίαρ Διάλεξη 4η 1

Perceptron s convergence 2

Proof of convergence Suppose that we have n training patterns that belong to two classes ω 1 and ω 2 the patterns of class ω 2 have been multiplied by -1 If the classes are linearly separable, the learning algorithm yields a solution weight vector w* with the property We can generalize using a non-negative threshold T

Proof of convergence Thus, the learning algorithm becomes We will consider only the indices k for which a correction takes place during training; thus

Proof of convergence Convergence of the algorithm means that, after some finite index k m The proof is as follows: Taking the inner product of w* with both sides Since each term is greater than T, then

Proof of convergence Taking the Cauchy-Schwartz inequality, results in Which may be written in the following form Substituting the nominator from previous inequality, we get

Proof of convergence Now, an alternative line of reasoning leads to a contradiction regarding w(k+1) 2 From the perceptron updating rule, we have or Having in mind that and letting we have that

Proof of convergence Adding these inequalities for j=1,2,,k yields Comparing this relation with the following which we derived earlier, i.e., we see that they establish conflicting bounds on w(k+1) 2 for sufficiently large k. In fact, k can be no larger than k m which is the solution to the equation: Therefore, k is finite. Q.E.D.

Activation functions 9

A single-input neuron The scalar input p is multiplied by the scalar weight w to form wp, one of the terms that is sent to the summer The other input, 1, is multiplied by a bias b and then passed to the summer (The bias is much like a weight, except that it has a constant input of 1. If you do not want to have a bias in a particular neuron, it can be omitted.) The summer output n, often referred to as the net input, goes into a transfer function f, which produces the scalar neuron output a. (We may also use the term activation function rather than transfer function and offset rather than bias.) The transfer function is chosen by the designer, and then the parameters w and b will be adjusted by some learning rule If we relate this simple model back to the biological neuron, the weight corresponds to the strength of a synapse, the cell body is represented by the summation and the transfer function, and the neuron output represents the signal on the axon

The hard limit transfer function The hard limit transfer function sets the output of the neuron to 0 if the function argument is less than 0, or 1 if its argument is greater than or equal to 0 We will use this function to create neurons that classify inputs into two distinct categories Observe the effect of the weight and the bias

The linear transfer function The output of a linear transfer function is equal to its input: Neurons with this transfer function are used in the ADALINE networks

The log-sigmoid transfer function The log-sigmoid transfer function takes the input (may have any value between + and - ) and outputs into the range 0 to 1, according to: It is commonly used in multilayer networks that are trained using backpropagation (in part because it is differentiable)

More transfer functions

More transfer functions

Multiple-input neuron The net input is n=w 1,1,p 1 + w 1,2 p 2 +... +w 1,R p R + b In matrix form: n=wp+b, where matrix W for the single neuron case has only one row The neuron output can be written as: α= f(wp+b) Notation: The first index indicates the particular neuron destination for that weight. The second index indicates the source of the signal fed to the neuron Thus, the indices in w 1,2 say that this weight represents the connection to the first (and only) neuron from the second source Abbreviated notation

A layer of neurons A single-layer network of S neurons each of the R inputs is connected to each of the neurons, and the weight matrix now has S rows You might ask if all the neurons in a layer must have the same transfer function. The answer is NO Abbreviated notation W=

Multiple layers of neurons Now consider a network with several layers. Each layer has its own weight matrix W, its own bias vector b, a net input vector n and an output vector a We may use superscripts to identify the layers. Specifically, we append the number of the layer as a superscript to the names for each of these variables

Multiple layers of neurons: Abbreviated notation

Perceptron for multiple classes Using the perceptron (training) rule to solve (during class lecture) the following classification problem: Use hardlim as the activation function and start with:

Perceptron for multiple classes Graphical illustration of the input

Perceptron for multiple classes Graphical illustration of the input

Perceptron for multiple classes The solution neural network is as follows:

Perceptron for multiple classes The solution decision boundaries