CMSC 421: Neural Computation. Applications of Neural Networks

Similar documents
Last update: October 26, Neural networks. CMSC 421: Section Dana Nau

AN INTRODUCTION TO NEURAL NETWORKS. Scott Kuindersma November 12, 2009

Neural networks. Chapter 19, Sections 1 5 1

Artificial Neural Networks

Unit 8: Introduction to neural networks. Perceptrons

Neural networks. Chapter 20. Chapter 20 1

Data Mining Part 5. Prediction

Neural networks. Chapter 20, Section 5 1

Neural Networks. Chapter 18, Section 7. TB Artificial Intelligence. Slides from AIMA 1/ 21

Lecture 4: Perceptrons and Multilayer Perceptrons

ARTIFICIAL INTELLIGENCE. Artificial Neural Networks

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Artificial Neural Networks

Artificial Neural Network

2015 Todd Neller. A.I.M.A. text figures 1995 Prentice Hall. Used by permission. Neural Networks. Todd W. Neller

AI Programming CS F-20 Neural Networks

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Artificial Neural Networks Examination, June 2005

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

100 inference steps doesn't seem like enough. Many neuron-like threshold switching units. Many weighted interconnections among units

Artificial Neural Networks Examination, June 2004

CS:4420 Artificial Intelligence

COMP-4360 Machine Learning Neural Networks

Artificial Intelligence

22c145-Fall 01: Neural Networks. Neural Networks. Readings: Chapter 19 of Russell & Norvig. Cesare Tinelli 1

Multilayer Neural Networks. (sometimes called Multilayer Perceptrons or MLPs)

Neural Networks biological neuron artificial neuron 1

Multilayer Neural Networks. (sometimes called Multilayer Perceptrons or MLPs)

Revision: Neural Network

Artifical Neural Networks

Artificial Neural Networks Examination, March 2004

Plan. Perceptron Linear discriminant. Associative memories Hopfield networks Chaotic networks. Multilayer perceptron Backpropagation

Incremental Stochastic Gradient Descent

Neural Networks: Introduction

Artificial Neural Networks. Edward Gatt

Lecture 7 Artificial neural networks: Supervised learning

Administration. Registration Hw3 is out. Lecture Captioning (Extra-Credit) Scribing lectures. Questions. Due on Thursday 10/6

Artificial Neural Networks

Machine Learning. Neural Networks

Artificial Neural Networks" and Nonparametric Methods" CMPSCI 383 Nov 17, 2011!

Neural Networks (Part 1) Goals for the lecture

Neural Networks. CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington

Part 8: Neural Networks

Introduction to Neural Networks

Course 395: Machine Learning - Lectures

Introduction To Artificial Neural Networks

Multilayer Neural Networks

4. Multilayer Perceptrons

Introduction to Machine Learning

Neural Networks Lecture 4: Radial Bases Function Networks

Lecture 5: Logistic Regression. Neural Networks

Introduction to Artificial Neural Networks

Artificial Neural Networks. Q550: Models in Cognitive Science Lecture 5

Ch.8 Neural Networks

COMP 551 Applied Machine Learning Lecture 14: Neural Networks

y(x n, w) t n 2. (1)

Introduction Neural Networks - Architecture Network Training Small Example - ZIP Codes Summary. Neural Networks - I. Henrik I Christensen

Artificial Neural Networks Examination, March 2002

SPSS, University of Texas at Arlington. Topics in Machine Learning-EE 5359 Neural Networks

CS 4700: Foundations of Artificial Intelligence

DEEP LEARNING AND NEURAL NETWORKS: BACKGROUND AND HISTORY

CSE 352 (AI) LECTURE NOTES Professor Anita Wasilewska. NEURAL NETWORKS Learning

Reification of Boolean Logic

Multilayer Perceptron

Neural Networks: Basics. Darrell Whitley Colorado State University

Machine Learning. Neural Networks. Le Song. CSE6740/CS7641/ISYE6740, Fall Lecture 7, September 11, 2012 Based on slides from Eric Xing, CMU

ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD

Introduction Biologically Motivated Crude Model Backpropagation

Neural Networks. Mark van Rossum. January 15, School of Informatics, University of Edinburgh 1 / 28

Introduction to feedforward neural networks

Multilayer Perceptrons (MLPs)

Artificial Neural Networks The Introduction

Artificial Neural Networks. Historical description

(Feed-Forward) Neural Networks Dr. Hajira Jabeen, Prof. Jens Lehmann

Sections 18.6 and 18.7 Artificial Neural Networks

Introduction to Natural Computation. Lecture 9. Multilayer Perceptrons and Backpropagation. Peter Lewis

Multilayer Perceptrons and Backpropagation

Grundlagen der Künstlichen Intelligenz

COMP304 Introduction to Neural Networks based on slides by:

Sections 18.6 and 18.7 Artificial Neural Networks

Neural Nets in PR. Pattern Recognition XII. Michal Haindl. Outline. Neural Nets in PR 2

NONLINEAR CLASSIFICATION AND REGRESSION. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

CN2 1: Introduction. Paul Gribble. Sep 10,

CSC242: Intro to AI. Lecture 21

Artificial Neural Network

Feedforward Neural Nets and Backpropagation

COMP9444 Neural Networks and Deep Learning 2. Perceptrons. COMP9444 c Alan Blair, 2017

Artificial Neural Networks D B M G. Data Base and Data Mining Group of Politecnico di Torino. Elena Baralis. Politecnico di Torino

Neural Networks with Applications to Vision and Language. Feedforward Networks. Marco Kuhlmann

Unit III. A Survey of Neural Network Model

Multilayer Neural Networks

Introduction to Neural Networks

Input layer. Weight matrix [ ] Output layer

Learning and Neural Networks

Deep Feedforward Networks

Chapter ML:VI (continued)

Artificial Neural Networks. Introduction to Computational Neuroscience Tambet Matiisen

Artificial Neural Network and Fuzzy Logic

Reading Group on Deep Learning Session 1

Neural Networks and the Back-propagation Algorithm

Transcription:

CMSC 42: Neural Computation definition synonyms neural networks artificial neural networks neural modeling connectionist models parallel distributed processing AI perspective Applications of Neural Networks pattern classification - virus/explosive detection, financial predictions, etc. image processing - character recognition, manufacturing inspection, etc. control - autonomous vehicles, industrial processes, etc. optimization - VLSI layout, scheduling, etc. bionics - prostheses, brain implants, etc. brain/cognitive models - memory, learning, disorders, etc.

Nature-Inspired Computation natural system biology, physics, etc. formal models, theories interdisciplinary neural networks genetic programming swarm intelligence self-replicating machines.. Inspiration for neural networks? applications computer science engineering nature Neural Networks model/theory random networks Hebbian learning perceptrons error backpropagation self-organizing maps applications pattern classification speech recognition image processing text-to-speech expert systems autonomous vehicles financial predictions associative memory data visualization + brain modeling 2

The Brain - complex - flow of information - what is known? - neurons - synapses (Purkinje cells; Golgi + Nissl stains) Neuron Information Processing 3

How Does the Brain Compute? A familiar example How fast is the processing? - cycle time vs. CPU - signal speeds How does it do that?! - massively parallel processing 0 neurons - different computational principles Summary: Brain as Inspiration network of neurons 0 neurons, 0 4 synapses neuron synapse flow of information spikes (pulses) Relevance to AI: Can a machine think? Alan Turing and weak AI Prospects for strong AI? 4

The Computer vs. The Brain information access control processing method how programmed adaptability global centralized sequential programmed minimally local decentralized massively parallel self-organizing prominently Our immediate focus: supervised/inductive learning History of Neural Networks 945-955: pre-computing 955-970: classical period 970-985: dark ages 985-995: renaissance 995-today: modern era perceptrons error backpropagation 5

Neural Computation basics feedforward networks - perceptrons - error backpropagation recurrent networks Neural Network Basics neural network = network + activation rule + learning rule 6

. network graph Neural Networks node/neuron activation level a i connection/synapse weight w ij a j w ij a i excitatory: w ij > 0 inhibitory: w ij < 0 j i + feedforward vs. recurrent networks 2. activation rule in i = " w ij a j a i = g(in i ) executing a neural network j r a. r w i a i Choices for Activation Function θ LTU (step) logistic (sigmoid) a i = step " (in i ) a i = " (in i ) local computations emergent behavior others: sign, tanh, linear, radial basis, 7

3. Learning Rule weight changes as function of local activity i w ij "w ij = f (a j,a i,in i,w ij,...) j Neural Computation basics feedforward networks - perceptrons - error backpropagation recurrent networks 8

Single Layer Networks o o o o o o o o number of layers? inputs outputs supervised learning: LMS Rule Perceptron Learning Rule derived using gradient descent associative memory, pattern classification input layer Elementary Perceptron LTU = linear threshold unit output r a. r w r a r {0, } r = response a r = step(in r ) θ r Example 0 3.0 5.0 -.0 θ r =.0 in r = 2.0 a r = 9

Perceptrons as Logic Gates Threshold needed to produce AND OR NOT a a 2 - θ r =.5 θ r = 0.5 θ r = -0.5 Linear separability: a 2 a 2 a 2 a a a Perceptron Learning Rule If the target output for unit r is t r w ri = w ri + "(t r # a r )a i η > 0 δ r = t r - a r Equivalent to the intuitive rules: If output is correct: don t change the weights If output is low (a r =0, t r =): If output is high (a r =, t r =0): Must also adjust threshold: increment weights for inputs = decrement weights for inputs = " r = " r # $(t r # a r ) (or, equivalently, assume there is a weight w r0 for an extra input unit that has a 0 =-: bias node) 0

Example of Perceptron Learning 0 3.0 5.0 -.0 θ r =.0 w ri = w ri + "(t r # a r )a i " r = " r # $(t r # a r ) Suppose η = 0. and t r = 0 + Perceptron Learning Algorithm repeatedly iterate through examples adjusting weights using perceptron learning rule until all outputs correct initialize the weights randomly or to all zero until outputs for all training examples are correct for each training example do compute the current output a r compare it to the target t r and update weights each pass through the training data is an epoch when will the algorithm terminate?

Perceptron Properties Perceptron Convergence Theorem: If there are a set of weights that are consistent with the training data (i.e., the data is linearly separable), the perceptron learning algorithm will converge on a solution. Perceptrons can only represent linear threshold functions and can therefore only learn functions that linearly separate the data, i.e., the positive and negative examples are separable by a hyperplane in n-dimensional space. Unfortunately, some functions (like xor) cannot be represented by a LTU. Error Backpropagation widely used neural network learning method seminal version about 960 (Rosenblatt) multiple versions since basic contemporary version popularized 985 uses multi-layer feedforward network 2

Uses Layered Feedforward Network O output units H hidden units I input units Representation Power of Multi-Layer Networks Theorem: any boolean function of N inputs can be represented by a network with one layer of hidden units. XOR a a 2 θ 3 =.5-2 a 3 a 4 θ 4 = 0.5 θ r = 0.5 a r = a 4 ~ a 3 = or(a,a 2 ) ~and(a,a 2 ) a r 3

Activation Function logistic (sigmoid) a r = " (in r ) Error Backpropagation Learning " k = (t k # a k )a k (# a k ) a k a j "w kj = #$ k a j $ ' " j = # w kj " k & ) a j (* a j ) % k ( "w ji = #$ j a i a i activity errors 4

Recall: Perceptron Learning Rule w ji = w ji + "(t j # a j )a i δ j Rewritten: "w ji = #$ j a i EBP Learning Rule " j = (t j # a j )a j (# a j ) output j w ji "w ji = #$ j a i hidden i output k s " k hidden j $ ' " j = # w kj " & k ) a j (* a j ) % ( k input i w ji "w ji = #$ j a i 5

Error Backpropagation repeatedly found to be effective in practice however, not guaranteed to find solution why? hill climbing can get stuck in local minima most widely used neural network method Error Backpropagation Applications NETtalk OCR ALVINN medical diagnosis plasma confinement neuroscience models cognitive models 6

NETtalk 26 (distributed representation) 60-20 (best: 20) 203 (7 x 29; local) typically > 8,000 weights training data: phonetic transcription of speech after 50 epochs: 95% correct on training data, 80% correct on test data window Optical Character Recognition (OCR) 7

ALVINN (Autonomous Land Vehicle In a Neural Network) neural net trained to drive a van along roads viewed through a TV camera speeds > 50 mph up to 20 miles (5 images/sec.) ALVINN s Neural Network Typical weights from a hidden node: H O I H - + - 4000 weights/biases 8