Artificial Neural Networks. Historical description

Similar documents
COMP9444 Neural Networks and Deep Learning 2. Perceptrons. COMP9444 c Alan Blair, 2017

Artifical Neural Networks

Introduction Biologically Motivated Crude Model Backpropagation

Introduction To Artificial Neural Networks

Artificial Neural Networks The Introduction

Introduction to Artificial Neural Networks

EEE 241: Linear Systems

Neural networks. Chapter 19, Sections 1 5 1

Lecture 7 Artificial neural networks: Supervised learning

Neural Networks. Chapter 18, Section 7. TB Artificial Intelligence. Slides from AIMA 1/ 21

Artificial Neural Networks

Simple Neural Nets For Pattern Classification

Artificial Intelligence

ARTIFICIAL INTELLIGENCE. Artificial Neural Networks

Lecture 4: Feed Forward Neural Networks

Neural Networks Introduction

Neural networks. Chapter 20. Chapter 20 1

Data Mining Part 5. Prediction

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Neural networks. Chapter 20, Section 5 1

Introduction to Neural Networks

Neural Networks and Fuzzy Logic Rajendra Dept.of CSE ASCET

CSC Neural Networks. Perceptron Learning Rule

Computational Intelligence Winter Term 2017/18

Feedforward Neural Nets and Backpropagation

2015 Todd Neller. A.I.M.A. text figures 1995 Prentice Hall. Used by permission. Neural Networks. Todd W. Neller

Unit III. A Survey of Neural Network Model

Computational Intelligence

Last update: October 26, Neural networks. CMSC 421: Section Dana Nau

Chapter 9: The Perceptron

Artificial Neural Networks. Q550: Models in Cognitive Science Lecture 5

Part 8: Neural Networks

Neural Networks (Part 1) Goals for the lecture

Artificial Neural Networks Examination, June 2005

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Artificial Neural Network

Supervised (BPL) verses Hybrid (RBF) Learning. By: Shahed Shahir

Artificial Neural Networks. Edward Gatt

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

ECE 471/571 - Lecture 17. Types of NN. History. Back Propagation. Recurrent (feedback during operation) Feedforward

Artificial neural networks

Revision: Neural Network

Plan. Perceptron Linear discriminant. Associative memories Hopfield networks Chaotic networks. Multilayer perceptron Backpropagation

Neural Networks: Introduction

CS4442/9542b Artificial Intelligence II prof. Olga Veksler. Lecture 5 Machine Learning. Neural Networks. Many presentation Ideas are due to Andrew NG

Artificial Neural Networks" and Nonparametric Methods" CMPSCI 383 Nov 17, 2011!

Lecture 4: Perceptrons and Multilayer Perceptrons

Introduction to Convolutional Neural Networks (CNNs)

Artificial Neural Network and Fuzzy Logic

CS:4420 Artificial Intelligence

Chapter ML:VI. VI. Neural Networks. Perceptron Learning Gradient Descent Multilayer Perceptron Radial Basis Functions

ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD

EE04 804(B) Soft Computing Ver. 1.2 Class 2. Neural Networks - I Feb 23, Sasidharan Sreedharan

Multilayer Perceptron

Machine Learning and Data Mining. Multi-layer Perceptrons & Neural Networks: Basics. Prof. Alexander Ihler

Introduction to Deep Learning

Introduction to Natural Computation. Lecture 9. Multilayer Perceptrons and Backpropagation. Peter Lewis

Administration. Registration Hw3 is out. Lecture Captioning (Extra-Credit) Scribing lectures. Questions. Due on Thursday 10/6

Supervised Learning. George Konidaris

Introduction to Neural Networks

How to do backpropagation in a brain

Deep Learning: a gentle introduction

CS 4700: Foundations of Artificial Intelligence

y(x n, w) t n 2. (1)

Artificial Neural Networks Examination, June 2004

Classification goals: Make 1 guess about the label (Top-1 error) Make 5 guesses about the label (Top-5 error) No Bounding Box

Neural Networks Lecturer: J. Matas Authors: J. Matas, B. Flach, O. Drbohlav

Fundamentals of Neural Network

NONLINEAR CLASSIFICATION AND REGRESSION. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

Fundamentals of Neural Networks

Introduction and Perceptron Learning

Artificial Neural Networks

Course 395: Machine Learning - Lectures

AN INTRODUCTION TO NEURAL NETWORKS. Scott Kuindersma November 12, 2009

Chapter 2 Single Layer Feedforward Networks

Perceptron. (c) Marcin Sydow. Summary. Perceptron

CS 4700: Foundations of Artificial Intelligence

An artificial neural networks (ANNs) model is a functional abstraction of the

DEEP LEARNING AND NEURAL NETWORKS: BACKGROUND AND HISTORY

(Feed-Forward) Neural Networks Dr. Hajira Jabeen, Prof. Jens Lehmann

22c145-Fall 01: Neural Networks. Neural Networks. Readings: Chapter 19 of Russell & Norvig. Cesare Tinelli 1

CMSC 421: Neural Computation. Applications of Neural Networks

Neural Networks. Fundamentals Framework for distributed processing Network topologies Training of ANN s Notation Perceptron Back Propagation

Multilayer Perceptron

Neural Networks. Nicholas Ruozzi University of Texas at Dallas

Apprentissage, réseaux de neurones et modèles graphiques (RCP209) Neural Networks and Deep Learning

Νεςπο-Ασαυήρ Υπολογιστική Neuro-Fuzzy Computing

Machine Learning. Neural Networks

Sections 18.6 and 18.7 Artificial Neural Networks

Multilayer Perceptrons and Backpropagation

Neural Networks. Learning and Computer Vision Prof. Olga Veksler CS9840. Lecture 10

AI Programming CS F-20 Neural Networks

Deep Neural Networks

Artificial neural networks

Supervised Learning in Neural Networks

ECE521 Lecture 7/8. Logistic Regression

Introduction to Neural Networks

A NEURAL FUZZY APPROACH TO MODELING THE THERMAL BEHAVIOR OF POWER TRANSFORMERS

Sections 18.6 and 18.7 Artificial Neural Networks

CSE446: Neural Networks Spring Many slides are adapted from Carlos Guestrin and Luke Zettlemoyer

Transcription:

Artificial Neural Networks Historical description Victor G. Lopez 1 / 23

Artificial Neural Networks (ANN) An artificial neural network is a computational model that attempts to emulate the functions of the brain. 2 / 23

Characteristics of ANNs Modern ANNs are complex arrangements of processing units able to adapt their parameters using learning techniques. Their plasticity, nonlinearity, robustness and highly distributed framework have attracted a lot of attention from many areas of research. Several applications have been studied using ANNs: classification, pattern recognition, clustering, function approximation, optimization, forecasting and prediction, among others. To date, ANN models are the artificial intelligence methods that imitate human intelligence more closely. 3 / 23

1890s: A neuron model Santiago Ramón y Cajal proposes the brain works in a parallel and distributed manner, with neurons as basic processing units. He described the first complete biological model of the neuron. 4 / 23

Neural synapses A neural synapse is the region where the axon of a neuron interacts with another neuron. A neuron usually receives information by means of its dendrites, but this is not always the case. Neurons share information using electrochemical signals. 5 / 23

Action Potential The signal sent by a single neuron is usually weak, but a neuron receives many inputs from many other neurons. The inputs from all the neurons are integrated. If a threshold is reached, the neuron sends a powerful signal through its axon, called an action potential. 6 / 23

Neural Pathways The action potential is an all-or-none signal. It doesn t matter if the threshold is barely reached or vastly surpassed, the resulting action potential is the same. This means that the action potential alone does not carry much information. All cerebral processes, like memory or learning, depend on neural pathways. There are over 10 11 neurons in the human brain, forming around 10 15 synapses. They form the basis of human intelligence and consciousness. 7 / 23

1943: Warren McCulloch (neurophysiologist) and Walter Pitts (matematician) wrote a paper describing a logical calculus of neural networks. Their model can, in principle, approximate any computable function. This is considered the birth of artificial intelligence. 8 / 23

1958: The Perceptron Frank Rosenblatt (psychologist) proposes the Perceptron with a novel method of supervised learning. This is the oldest neural network still in use today. 9 / 23

Single-neuron Perceptron Here, the activation function f was selected as a saturation function. This simulates the all-or-none property of the action potential. The single-neuron Perceptron can solve classification problems of two linearly separable groups. 10 / 23

Perceptron with a Layer of Neurons Using several neurons, the Perceptron can classify objects into many categories, as long as they are linearly separable. The number of total categories is 2 S, with S the number of neurons. 11 / 23

Training algorithm per neuron 1 Initialize the weights W 0. 2 Compute the output of the network for input p k. If the output is correct, set W k+1 = W k 3 If the output is incorrect, set W k+1 = W k ηp k, if W T k p k 0 W k+1 = W k + ηp k, if W T k p k < 0 Here, 0 < η 1 is the learning rate. 12 / 23

Logical gates AND and OR Separation of the outputs of the logical gates AND and OR are simple examples of problems solvable by the single-layer Perceptron. In contrast, the outputs of the XOR gate are not linearly separable. x 1 x 2 y 0 0 0 0 1 0 1 0 0 1 1 1 AND x 1 x 2 y 0 0 0 0 1 1 1 0 1 1 1 1 OR x 1 x 2 y 0 0 0 0 1 1 1 0 1 1 1 0 XOR 13 / 23

1959: ADALINE and MADALINE Bernard Widrow and Marcian Hoff developed models called ADALINE (adaptive linear elements) and MADALINE (Multiple ADALINE). The main difference with respect to the Perceptron is the absence of the threshold activation function. Training of these networks is performed using derivatives. 14 / 23

1970s: in ANNs research After the successful introduction and development of ANNs during the 1960s, interest in their applications decayed for almost two decades. The limitations of the single-layer Perceptron narrowed its possible practical implementations. Theoretical research showed that a multilayer Perceptron would drastically improve its performance, but there was no training algorithm for it. 15 / 23

1986: 16 / 23

1986: In his 1974 PhD thesis, Paul Werbos proposed to use the backpropagation algorithm as a solution to the multilayer Perceptron training problem. His suggestion, however, remained ignored for more than a decade. In 1986, the backpropagation method is finally popularized in a paper by Rumelhart, Hinton and Williams. The multilayer Perceptron became the most powerful ANN model to date. It is proven to solve nonlinearly separable classification problems, it can approximate any continuous function, it generalizes from particular samples, among many other applications. 17 / 23

Backpropagation Training This is a supervised learning. Then, we have a list of inputs and their corresponding target outputs, (p k, t k ). We can compute the output of the NN for each given input p k. This is called the forward propagation step. For a 3 layer network, this would be a k = f 3 (W 3 f 2 (W 2 f 1 (W 1 p k + b 1 ) + b 2 ) + b 3 ) Define the output error as e k = t k a k. Now, it is our interest to minimize the squared error J = 1 2 e2 k = 1 2 (t k a k ) 2 or the average sum of the squared error J = 1 2Q Q k=1 e 2 k = 1 2Q Q (t k a k ) 2 k=1 18 / 23

Backpropagation Training Use a gradient descent algorithm to update the weights W k while minimizing the error e k W k+1 = W k + W k W k = η J W k The chain rule for derivatives can be used to obtain a clearer expression J W k = J e k e k a k a k W k From the previous definitions we note that J e k = e k, e k a k = 1 and a k W k depends on the activation functions f i. Notice that all activation functions must be differentiable. 19 / 23

Backpropagation Algorithm 1 Initialize the weights W 0. 2 By forward propagation, get a k. 3 Calculate the error e k = t k a k. 4 Update the neural weights as W k+1 = W k + η a k W k e k 20 / 23

Activation functions 21 / 23

Late 1990s - Early 2000 s: in ANNs research In the 1990s, several applications of ANNs are studied and implemented. Areas as vision, pattern recognition, unsupervised learning and reinforcement learning take advantage of the ANNs adaptive characteristics. Late in that decade, a new difficulty delays the advancement in the field. The basic backpropagation algorithm is not appropriate for several hidden layers, mainly because of limited computational capabilities. Many researchers became pessimistic about ANNs. 22 / 23

2006-2016: Deep learning In 2006, Hinton, Osindero and Teh published a fast learning algorithm for deep belief networks. This marks the dawn of deep learning. The decade of 2010s has seen a boom in deep neural network applications. Companies as Microsoft, Google and Facebook have developed advanced deep learning ANNs. Optimism has returned to the field and human-level intelligence is expected to be achieved in a few decades. 23 / 23