EPL442: Computational

Similar documents
COGS Q250 Fall Homework 7: Learning in Neural Networks Due: 9:00am, Friday 2nd November.

Neural Networks biological neuron artificial neuron 1

Introduction to Machine Learning

COMP 551 Applied Machine Learning Lecture 14: Neural Networks

Multilayer Neural Networks. (sometimes called Multilayer Perceptrons or MLPs)

Comments. Assignment 3 code released. Thought questions 3 due this week. Mini-project: hopefully you have started. implement classification algorithms

Multilayer Neural Networks. (sometimes called Multilayer Perceptrons or MLPs)

Neural Nets Supervised learning

Serious limitations of (single-layer) perceptrons: Cannot learn non-linearly separable tasks. Cannot approximate (learn) non-linear functions

Neural Networks: Basics. Darrell Whitley Colorado State University

4. Multilayer Perceptrons

Artifical Neural Networks

Introduction to Natural Computation. Lecture 9. Multilayer Perceptrons and Backpropagation. Peter Lewis

Multilayer Perceptrons and Backpropagation

Artificial Intelligence

C4 Phenomenological Modeling - Regression & Neural Networks : Computational Modeling and Simulation Instructor: Linwei Wang

Unit 8: Introduction to neural networks. Perceptrons

ECLT 5810 Classification Neural Networks. Reference: Data Mining: Concepts and Techniques By J. Hand, M. Kamber, and J. Pei, Morgan Kaufmann

AI Programming CS F-20 Neural Networks

Backpropagation Neural Net

y(x n, w) t n 2. (1)

Deep Learning & Artificial Intelligence WS 2018/2019

epochs epochs

Introduction Neural Networks - Architecture Network Training Small Example - ZIP Codes Summary. Neural Networks - I. Henrik I Christensen

AN INTRODUCTION TO NEURAL NETWORKS. Scott Kuindersma November 12, 2009

Artificial Neural Networks

Data Mining Part 5. Prediction

Unit III. A Survey of Neural Network Model

Introduction to Neural Networks

Neural Networks. Yan Shao Department of Linguistics and Philology, Uppsala University 7 December 2016

Mark Gales October y (x) x 1. x 2 y (x) Inputs. Outputs. x d. y (x) Second Output layer layer. layer.

The error-backpropagation algorithm is one of the most important and widely used (and some would say wildly used) learning techniques for neural

Training Multi-Layer Neural Networks. - the Back-Propagation Method. (c) Marcin Sydow

2015 Todd Neller. A.I.M.A. text figures 1995 Prentice Hall. Used by permission. Neural Networks. Todd W. Neller

Last update: October 26, Neural networks. CMSC 421: Section Dana Nau

Neural networks. Chapter 19, Sections 1 5 1

Computational statistics

Neural Networks. Chapter 18, Section 7. TB Artificial Intelligence. Slides from AIMA 1/ 21

CMSC 421: Neural Computation. Applications of Neural Networks

Feedforward Neural Nets and Backpropagation

Pattern Recognition Prof. P. S. Sastry Department of Electronics and Communication Engineering Indian Institute of Science, Bangalore

22c145-Fall 01: Neural Networks. Neural Networks. Readings: Chapter 19 of Russell & Norvig. Cesare Tinelli 1

Supervised Learning in Neural Networks

Artificial Neural Networks Examination, March 2004

ECE521 Lectures 9 Fully Connected Neural Networks

Neural Networks. Nicholas Ruozzi University of Texas at Dallas

Neural Networks, Computation Graphs. CMSC 470 Marine Carpuat

Deep Feedforward Networks

Machine Learning: Multi Layer Perceptrons

Neural Networks and Deep Learning

Machine Learning for Large-Scale Data Analysis and Decision Making A. Neural Networks Week #6

Artificial Neural Network

Artificial Neural Networks Examination, June 2005

Artificial Neural Networks

Speaker Representation and Verification Part II. by Vasileios Vasilakakis

(Feed-Forward) Neural Networks Dr. Hajira Jabeen, Prof. Jens Lehmann

Neural Networks. Xiaojin Zhu Computer Sciences Department University of Wisconsin, Madison. slide 1

Lecture 13 Back-propagation

Lecture 5: Logistic Regression. Neural Networks

Neural Networks. CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

EEE 241: Linear Systems

Neuro-Fuzzy Comp. Ch. 4 March 24, R p

CSE446: Neural Networks Spring Many slides are adapted from Carlos Guestrin and Luke Zettlemoyer

CS:4420 Artificial Intelligence

CSCI567 Machine Learning (Fall 2018)

Multilayer Perceptrons (MLPs)

Multilayer Perceptron

Neural networks. Chapter 20. Chapter 20 1

Classification goals: Make 1 guess about the label (Top-1 error) Make 5 guesses about the label (Top-5 error) No Bounding Box

SPSS, University of Texas at Arlington. Topics in Machine Learning-EE 5359 Neural Networks

CS 6501: Deep Learning for Computer Graphics. Basics of Neural Networks. Connelly Barnes

Artificial Neural Networks. Edward Gatt

6.034f Neural Net Notes October 28, 2010

Feed-forward Networks Network Training Error Backpropagation Applications. Neural Networks. Oliver Schulte - CMPT 726. Bishop PRML Ch.

A thorough derivation of back-propagation for people who really want to understand it by: Mike Gashler, September 2010

Retrieval of Cloud Top Pressure

Introduction to Neural Networks

Multilayer Perceptron Tutorial

CSE 352 (AI) LECTURE NOTES Professor Anita Wasilewska. NEURAL NETWORKS Learning

Artificial Neural Networks

Jakub Hajic Artificial Intelligence Seminar I

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Machine Learning

CSC 578 Neural Networks and Deep Learning

Deep Feedforward Networks

Course 395: Machine Learning - Lectures

Neural networks. Chapter 20, Section 5 1

CSE 190 Fall 2015 Midterm DO NOT TURN THIS PAGE UNTIL YOU ARE TOLD TO START!!!!

Input layer. Weight matrix [ ] Output layer

Classification with Perceptrons. Reading:

Rapid Introduction to Machine Learning/ Deep Learning

MIDTERM: CS 6375 INSTRUCTOR: VIBHAV GOGATE October,

Neural Networks and the Back-propagation Algorithm

Neural Networks. Bishop PRML Ch. 5. Alireza Ghane. Feed-forward Networks Network Training Error Backpropagation Applications

Neural Network Tutorial & Application in Nuclear Physics. Weiguang Jiang ( 蒋炜光 ) UTK / ORNL

Deep Neural Networks (3) Computational Graphs, Learning Algorithms, Initialisation

Artificial Neural Networks Examination, June 2004

Data Mining & Machine Learning

Introduction to feedforward neural networks

Transcription:

EPL442: Computational Learning Systems Lab 2 Vassilis Vassiliades Department of Computer Science University of Cyprus

Outline Artificial Neuron Feedforward Neural Network Back-propagation Algorithm Notes Assignment Questions? Demos Cortex Pro ALVINN

Artificial Neuron Summing Junction Output * W X Inputs *W 2 Y Σ * W 3 Activation Function (e.g., sigmoid) Weights (incl. bias) Y a* X * W X * W W e 2 2 3

Feedforward Neural Network IN H H2 OUT IN 2 H 2 H2 2 OUT 2 Input Layer (inactive - meaning no computation) Hidden Layer Hidden Layer 2 Output Layer

Back-propagation Algorithm (Online Updating) Step by Step

Pattern p is presented f X

Forward Pass w( bias ) w ( x ) f X w( x 2) y f ( w w x w x ) ( bias ) ( x) ( x2) 2

Forward Pass w( bias )2 f X w( x )2 w( x 2)2 y f ( w w x w x ) 2 2 ( bias )2 ( x)2 ( x2)2 2

Forward Pass w ( bias )3 w 3 f X w 23 y f ( w w y w y ) 3 3 ( bias )3 3 23 2

Forward Pass w( bias )4 X f w 4 w 24 y f ( w w y w y ) 4 4 ( bias )4 4 24 2

Forward Pass w ( bias )5 f X w 35 w 45 y f ( w w y w y ) 5 5 ( bias )5 35 3 45 4

Backward Pass - Stage Derivative of the error with respect to the output (y 5 ) Derivative of the output node s activation function (i.e. sigmoid) y ( y ) ( y t ) 5 5 5 5 X f t δ δ 5 y 5 Target output (in our case we have only output) Actual output Assuming we are minimising the sum-of-squares error function and that each neuron has a sigmoid activation function with slope α=

Backward Pass - Stage y ( y ) ( w ) 3 3 3 35 5 δ 3 X f δ δ5 w 35

Backward Pass - Stage δ 3 X f δ 5 δ 4 w 45 y ( y ) ( w ) 4 4 4 45 5

Backward Pass - Stage δ δ 3 X w 3 f w 4 δ 4 δ 5 y ( y ) ( w w ) 3 3 4 4

Backward Pass - Stage δ δ 3 X f δ 5 δ 2 δ 4 w 23 w 24 y ( y ) ( w w ) 2 2 2 23 3 24 4

Backward Pass - Stage 2 w' ( bias ) δ δ 3 X w' x ( x ) f δ 5 w' x δ 2 δ 4 2 4 '( 2) w ' ( bias ) w ( bias ) w' w x ( x) ( x) w w x ' ( x 2) ( x 2) 2

Backward Pass - Stage 2 δ 3 w' ( bias )2 X f δ 5 w' ( x )2 δ 2 δ 4 w w' w ( x2)2 ' ( bias )2 w ( bias )2 2 w' w x ( x)2 ( x)2 2 w w x ' ( x 2)2 ( x 2)2 2 2

Backward Pass - Stage 2 w' ( bias )3 δ 3 X w' 3 f δ 5 w' δ 23 4 w w ' ( bias )3 ( bias )3 3 w ' w y 3 3 3 w' w y 23 23 3 2

Backward Pass - Stage 2 w' ( bias )4 X f w' 4 δ 4 δ 5 w' 24 w w ' ( bias )4 ( bias )4 4 w ' w y 4 4 4 w' w y 24 24 4 2

Backward Pass - Stage 2 w' ( bias )5 X f w' 35 δ 5 w' 45 w' w ( bias)5 ( bias)5 5 w' w y w w y 35 35 5 3 ' 45 45 5 4

Pattern p+ is presented f X The procedure is the same as before

Notes In backward pass you could update the weights immediately after you calculate the deltas. Here the backward pass was done in 2 distinct stages: a) propagation of errors backwards in order to evaluate the derivatives and b) weight adjustment using the calculated derivatives. In class notes the derivative of the error is (target output). If you prefer it like this then you need to put a plus sign (+) in front of the learning rate in the weight update equations (as in class notes). A large learning rate is equivalent to big changes in the weights, thus large jumps in the weight space. This is not always desirable.

Notes To minimise the occurences of local minima: Change the learning rate (either start with a large value and progressively decrease it or intelligently adapt it) Add more hidden nodes (be careful of the overfitting problem) Add momentum to the weight update equations Add noise Overfitting occurs when the neural network: is trained for too long (to avoid this problem stop training early) has a lot of hidden nodes (to avoid this problem do model selection) In the batch update version the weight changes are accumulated and applied after each epoch instead of after each pattern. Weights should be initialised to small random values in the range [-,] and input data should be normalised in the range [0,].

Assignment Assignment Document: www.cs.ucy.ac.cy/~vvassi0/epl442 Submission deadline: 9 October 2009 Deliverables Report in pdf format max. 3 pages Source code with comments Other files (training.dat, test.dat, parameters.dat etc.) All in a zip file to: v.vassiliades@cs.ucy.ac.cy

Assignment - Grading Completeness of the deliverables All written in the assignment description Source code correctness Automated tests Interactive tests Source code quality Comments, design Quality of the report Time of submission

Questions?

Demos Cortex Pro www.tech.plym.ac.uk/soc/staff/guidbugm/software.htm Download and experiment with. ALVINN (Autonomous Land Vehicle In a Neural Network) See video