Artificial Intelligence

Similar documents
Revision: Neural Network

EEE 241: Linear Systems

Data Mining Part 5. Prediction

ARTIFICIAL INTELLIGENCE. Artificial Neural Networks

(Feed-Forward) Neural Networks Dr. Hajira Jabeen, Prof. Jens Lehmann

Artificial Neural Networks" and Nonparametric Methods" CMPSCI 383 Nov 17, 2011!

CSC242: Intro to AI. Lecture 21

COMP 551 Applied Machine Learning Lecture 14: Neural Networks

Feedforward Neural Nets and Backpropagation

Artificial Neural Networks

18.6 Regression and Classification with Linear Models

Artificial Neural Networks. Historical description

CS 6501: Deep Learning for Computer Graphics. Basics of Neural Networks. Connelly Barnes

Introduction to Natural Computation. Lecture 9. Multilayer Perceptrons and Backpropagation. Peter Lewis

Neural Networks, Computation Graphs. CMSC 470 Marine Carpuat

Artificial Neural Networks. Edward Gatt

ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD

Neural Networks and the Back-propagation Algorithm

Machine Learning for Large-Scale Data Analysis and Decision Making A. Neural Networks Week #6

Introduction to Neural Networks

2015 Todd Neller. A.I.M.A. text figures 1995 Prentice Hall. Used by permission. Neural Networks. Todd W. Neller

AI Programming CS F-20 Neural Networks

Neural Networks. CSE 6363 Machine Learning Vassilis Athitsos Computer Science and Engineering Department University of Texas at Arlington

Lab 5: 16 th April Exercises on Neural Networks

Neural networks. Chapter 19, Sections 1 5 1

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Neural Networks and Fuzzy Logic Rajendra Dept.of CSE ASCET

AN INTRODUCTION TO NEURAL NETWORKS. Scott Kuindersma November 12, 2009

Machine Learning. Neural Networks

Neural networks. Chapter 20. Chapter 20 1

CSC321 Lecture 5: Multilayer Perceptrons

Machine Learning and Data Mining. Multi-layer Perceptrons & Neural Networks: Basics. Prof. Alexander Ihler

CSE 352 (AI) LECTURE NOTES Professor Anita Wasilewska. NEURAL NETWORKS Learning

Artificial Neural Networks. Q550: Models in Cognitive Science Lecture 5

y(x n, w) t n 2. (1)

Multilayer Perceptron

Artificial Neural Networks. MGS Lecture 2

Artificial Neural Network and Fuzzy Logic

CMSC 421: Neural Computation. Applications of Neural Networks

Neural Networks. Bishop PRML Ch. 5. Alireza Ghane. Feed-forward Networks Network Training Error Backpropagation Applications

Unit III. A Survey of Neural Network Model

COGS Q250 Fall Homework 7: Learning in Neural Networks Due: 9:00am, Friday 2nd November.

Artificial Neural Networks Examination, June 2005

Artifical Neural Networks

Linear classification with logistic regression

Lecture 4: Perceptrons and Multilayer Perceptrons

ECLT 5810 Classification Neural Networks. Reference: Data Mining: Concepts and Techniques By J. Hand, M. Kamber, and J. Pei, Morgan Kaufmann

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Neural Networks. Chapter 18, Section 7. TB Artificial Intelligence. Slides from AIMA 1/ 21

Recurrent Neural Networks

SPSS, University of Texas at Arlington. Topics in Machine Learning-EE 5359 Neural Networks

Artificial Neural Networks

Lecture 4: Feed Forward Neural Networks

Introduction to Artificial Neural Networks

Neural Networks: Introduction

4. Multilayer Perceptrons

Neural Networks and Deep Learning

CSE 190 Fall 2015 Midterm DO NOT TURN THIS PAGE UNTIL YOU ARE TOLD TO START!!!!

Artificial Neural Networks. Introduction to Computational Neuroscience Tambet Matiisen

Reading Group on Deep Learning Session 1

CSC 411 Lecture 10: Neural Networks

Nonlinear Classification

Multilayer Perceptrons and Backpropagation

Neural Networks. Nicholas Ruozzi University of Texas at Dallas

Course 395: Machine Learning - Lectures

Unit 8: Introduction to neural networks. Perceptrons

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Neural Networks. Intro to AI Bert Huang Virginia Tech

Serious limitations of (single-layer) perceptrons: Cannot learn non-linearly separable tasks. Cannot approximate (learn) non-linear functions

Introduction Biologically Motivated Crude Model Backpropagation

Introduction to Deep Learning

Computational statistics

Learning and Neural Networks

Sections 18.6 and 18.7 Analysis of Artificial Neural Networks

CSC321 Lecture 5 Learning in a Single Neuron

Neural Networks. Mark van Rossum. January 15, School of Informatics, University of Edinburgh 1 / 28

Sections 18.6 and 18.7 Artificial Neural Networks

ECE 471/571 - Lecture 17. Types of NN. History. Back Propagation. Recurrent (feedback during operation) Feedforward

ECE521 Lectures 9 Fully Connected Neural Networks

HISTORY. In the 70 s vs today variants. (Multi-layer) Feed-forward (perceptron) Hebbian Recurrent

Artificial Neural Network

Last update: October 26, Neural networks. CMSC 421: Section Dana Nau

ARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92

Introduction to Machine Learning Spring 2018 Note Neural Networks

Introduction Neural Networks - Architecture Network Training Small Example - ZIP Codes Summary. Neural Networks - I. Henrik I Christensen

Supervised (BPL) verses Hybrid (RBF) Learning. By: Shahed Shahir

Part 8: Neural Networks

Statistical NLP for the Web

CS 4700: Foundations of Artificial Intelligence

Neural networks. Chapter 20, Section 5 1

Sections 18.6 and 18.7 Artificial Neural Networks

Lecture 6. Notes on Linear Algebra. Perceptron

Logistic Regression & Neural Networks

Artificial neural networks

From perceptrons to word embeddings. Simon Šuster University of Groningen

Neural Networks Lecturer: J. Matas Authors: J. Matas, B. Flach, O. Drbohlav

Intro to Neural Networks and Deep Learning

Multilayer Neural Networks. (sometimes called Multilayer Perceptrons or MLPs)

Lecture 17: Neural Networks and Deep Learning

Introduction and Perceptron Learning

Transcription:

Artificial Intelligence Jeff Clune Assistant Professor Evolving Artificial Intelligence Laboratory

Announcements Be making progress on your projects!

Three Types of Learning Unsupervised Supervised Reinforcement

Perceptron Learning Rule Typically applied one example at a time choosing samples at random Guaranteed to find a perfect boundary if data are linearly separable Example Run

Perceptron Learning Rule Not guaranteed to find optimal solution if data are not linearly separable below it gets to the minimal error many times, but keeps changing the weights With Non-Separable Data

Perceptron Learning Rule Not guaranteed to find optimal solution if data are not linearly separable below it gets to the minimal error many times, but keeps changing the weights decreasing learning rate (setting alpha to 1/numIterations) will find a minimum-error solution With Non-Separable Data With Cooling

Perceptron Learning Rule Lots!,, With Linearly Separable Data With Non-Separable Data With Cooling

Logistic Regression in contrast to linear regression Threshold ( step-function ) has drawbacks non-differentiable always produces a 1 or 0 classification - even near the boundary

Logistic Regression Threshold ( step-function ) has drawbacks non-differentiable always produces a 1 or 0 classification - even near the boundary Solutions the differentiable logistic function output is a probability of the class - 0.5 near boundary towards 0 or 1 farther away

Logistic Regression Sigmoid for Earthquakes vs. Explosions

No closed-form solution Logistic Regression Weight update rule (see book for derivation) Linearly Separable Data Bit slower than linear regression, but more predictable not linearly separable, fixed learning rate Faster than linear regression not linearly separable, decaying learning rate Faster than linear regression

Logistic Regression

Your brain is a neural network Neural Networks

Neural Networks We create computational brains called Artificial Neural Networks or ANNs

Neural Networks Nodes feed forward - layered

Neural Networks Nodes feedforward - layers recurrent - allows memory recurrent

Neural Networks Connections excitatory or inhibitory different weights (magnitudes) typically real valued

Neural Networks Nodes often layered function of inputs - step/threshold - sigmoid - other...

Neural Networks Nodes often layered function of inputs - step/threshold - sigmoid - other... biases - often treated as another input set to 1 - always use them! - otherwise decision boundary has to pass through origin - also good when default output is non-zero, such as a robot that moves forward unless X - make sure they are on for your projects!

Neural Networks Simple example: kiwi classifier binary inputs weights =?? bias = 1 Roundess Furriness T=0 = 1 0 f(x) = 1 if >=0, Bias 0 otherwise

Non-Linear Activation Functions Combining many of them builds an ANN that can represent non-linear functions - because activation functions are nonlinear

Outputs Can have single or multiple outputs e.g. one output for each class

Perceptron Perceptron: Single unit Default activation function: step Sigmoid: sigmoid perceptron Perceptron Network: single-layer network McCullough & Pitts (1943)

Perceptron Training Rule Training: step activation function: perceptron learning rule logistic activation function: gradient descent both work for linearly separable data - we already new that just a new metaphor now

Neural Networks Outputs 1 if >= number in node 1 neuron2 (right) 1 2D Non-Linear Mapping 1 Outputs of each node x y Left Center Right Output 1 1 1 1 1 0 1 0 1 0 0 1 0 1 0 0 1 1 0 0 0 0 0 0 neuron3 (center) 3D neuron1 (left)

Neural Networks neuron2 (right) 1 1 neuron3 (center) neuron1 (left) 3D

Neural Networks Cover s theorem: The probability that classes are linearly separable increases when the features are nonlinearly mapped to a higher dimensional feature space. [Coover 1965] The output layer requires linear separability. The purpose of the hidden layers is to make the problem linearly separable! From: http://130.236.96.13/edu/courses/tbmi26/pdfs/lectures/le5.pdf

Neural Networks Multi-layer networks thus allow non-linear regression

Neural Networks Multi-layer networks thus allow non-linear regression Single hidden layer (often very large): can represent any continuous function Two hidden layers: can represent any discontinuous function

Neural Networks Multi-layer networks thus allow non-linear regression Single hidden layer (often very large): can represent any continuous function Two hidden layers: can represent any discontinuous function But how do we train them?

Training Multi-Layer Neural Networks General Idea: Propagate the error backwards Called Backpropagation

Backpropagation Can still write the equation of a neuron Activation of Neuron 5 Bias

Backpropagation Can still write the equation of a neuron Allows us to calculate loss and take derivative Activation of Neuron 5 Bias

Backpropagation Can still write the equation of a neuron Allows us to calculate loss and take derivative At the output, the loss eq. is the same as before prev. gradient descent eq. now generalized to multiple output nodes

Backpropagation But what is the error at hidden nodes? We don t know what the right answer is for hidden nodes

Backpropagation

Backpropagation Detailed Pseudocode in Book

ANNs How do we choose the topology? number of layers number of neurons per layer Too many neurons = memorization/overfitting No perfect answer try different ones and cross-validate evolve (take my class in the Spring!) many others

Deep Learning Hierarchically composed feature representations Deep Neural Networks ~6+ Layers Each layer transforms data to new space could be higher-dimensional, lower, or just different

Hierarchically composed feature representations

Learning features relevant to the data Lehman, Clune, & Risi 2015 Lee et al. 2007

Deep Learning Follows biology closely V1, V2, etc. in Visual Cortex go from edges to more abstract features eventually to a Halle Berry Neuron responded to her picture a line drawing of her her picture in her Catwoman costume her name (typed out) Quiroga et al. Nature 2005 For more, see: http://goo.gl/zcrgic