Computational Intelligence Winter Term 2009/10

Similar documents
Computational Intelligence

Computational Intelligence

Simple Neural Nets for Pattern Classification: McCulloch-Pitts Threshold Logic CS 5870

COMP9444 Neural Networks and Deep Learning 2. Perceptrons. COMP9444 c Alan Blair, 2017

Neural Networks. Chapter 18, Section 7. TB Artificial Intelligence. Slides from AIMA 1/ 21

Computational Intelligence Winter Term 2017/18

Artificial Neural Networks The Introduction

Computational Intelligence

ARTIFICIAL INTELLIGENCE. Artificial Neural Networks

Neural Networks. Fundamentals of Neural Networks : Architectures, Algorithms and Applications. L, Fausett, 1994

Artificial Neural Network and Fuzzy Logic

Neural Networks Introduction CIS 32

Neural networks. Chapter 20, Section 5 1

Introduction Biologically Motivated Crude Model Backpropagation

Neural networks. Chapter 19, Sections 1 5 1

Neural Networks Introduction

Last update: October 26, Neural networks. CMSC 421: Section Dana Nau

Artificial neural networks

Lecture 4: Feed Forward Neural Networks

Artificial Neural Networks. Part 2

Artificial Neural Networks

Introduction and Perceptron Learning

EE04 804(B) Soft Computing Ver. 1.2 Class 2. Neural Networks - I Feb 23, Sasidharan Sreedharan

Neural networks. Chapter 20. Chapter 20 1

Artificial Neural Networks. Historical description

EEE 241: Linear Systems

Artificial Neural Networks

Master Recherche IAC TC2: Apprentissage Statistique & Optimisation

Multi-Dimensional Neural Networks: Unified Theory

Neural Networks: Introduction

Feedforward Neural Nets and Backpropagation

Data Mining Part 5. Prediction

CS:4420 Artificial Intelligence

Introduction to Artificial Neural Networks

CSC321 Lecture 5: Multilayer Perceptrons

Introduction To Artificial Neural Networks

CE213 Artificial Intelligence Lecture 13

Sections 18.6 and 18.7 Artificial Neural Networks

Artificial Neural Networks. Q550: Models in Cognitive Science Lecture 5

Neural Networks for Machine Learning. Lecture 2a An overview of the main types of neural network architecture

Course 395: Machine Learning - Lectures

Intelligent Systems Discriminative Learning, Neural Networks

Perceptron. (c) Marcin Sydow. Summary. Perceptron

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Machine Learning. Neural Networks

Computational Intelligence

Artificial Neural Network

Simple Neural Nets For Pattern Classification

Artificial Neural Networks. MGS Lecture 2

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

COMP304 Introduction to Neural Networks based on slides by:

Lecture 7 Artificial neural networks: Supervised learning

Neural Networks. Xiaojin Zhu Computer Sciences Department University of Wisconsin, Madison. slide 1

Week 3: More Processing, Bits and Bytes. Blinky, Logic Gates, Digital Representations, Huffman Coding. Natural Language and Dialogue Systems Lab

Symbolic vs. subsymbolic representation in cognitive science and artificial intelligence Vladimír Kvasnička FIIT STU.

22c145-Fall 01: Neural Networks. Neural Networks. Readings: Chapter 19 of Russell & Norvig. Cesare Tinelli 1

Artifical Neural Networks

CS 4700: Foundations of Artificial Intelligence

Sections 18.6 and 18.7 Artificial Neural Networks

2015 Todd Neller. A.I.M.A. text figures 1995 Prentice Hall. Used by permission. Neural Networks. Todd W. Neller

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD

Neural Networks and Fuzzy Logic Rajendra Dept.of CSE ASCET

Linear Regression, Neural Networks, etc.

Introduction to Neural Networks

Machine Learning. Neural Networks. Le Song. CSE6740/CS7641/ISYE6740, Fall Lecture 7, September 11, 2012 Based on slides from Eric Xing, CMU

Lecture 4: Perceptrons and Multilayer Perceptrons

Νεςπο-Ασαυήρ Υπολογιστική Neuro-Fuzzy Computing

PETER PAZMANY CATHOLIC UNIVERSITY Consortium members SEMMELWEIS UNIVERSITY, DIALOG CAMPUS PUBLISHER

Neural Networks Lecturer: J. Matas Authors: J. Matas, B. Flach, O. Drbohlav

Grundlagen der Künstlichen Intelligenz

Lab 5: 16 th April Exercises on Neural Networks

Networks of McCulloch-Pitts Neurons

CS 4700: Foundations of Artificial Intelligence

New Delhi & Affiliated to VTU, Belgavi ) Oorgaum, Kolar Dist ,Karnataka

Input layer. Weight matrix [ ] Output layer

Artificial Neural Networks" and Nonparametric Methods" CMPSCI 383 Nov 17, 2011!

AN INTRODUCTION TO NEURAL NETWORKS. Scott Kuindersma November 12, 2009

Machine Learning (CSE 446): Neural Networks

CS 256: Neural Computation Lecture Notes

Chapter ML:VI. VI. Neural Networks. Perceptron Learning Gradient Descent Multilayer Perceptron Radial Basis Functions

Multi-layer Perceptron Networks

17 Neural Networks NEURAL NETWORKS. x XOR 1. x Jonathan Richard Shewchuk

Simulating Neural Networks. Lawrence Ward P465A

Neural Networks. Fundamentals Framework for distributed processing Network topologies Training of ANN s Notation Perceptron Back Propagation

Artificial Neural Networks Lecture Notes Part 3

Multilayer Perceptron Tutorial

Neural Networks (Part 1) Goals for the lecture

Compartmental Modelling

Multilayer Perceptron

Fundamentals of Neural Networks

Computational Intelligence Winter Term 2018/19

CSE446: Neural Networks Spring Many slides are adapted from Carlos Guestrin and Luke Zettlemoyer

Vote. Vote on timing for night section: Option 1 (what we have now) Option 2. Lecture, 6:10-7:50 25 minute dinner break Tutorial, 8:15-9

(Feed-Forward) Neural Networks Dr. Hajira Jabeen, Prof. Jens Lehmann

In the Name of God. Lecture 9: ANN Architectures

Principles of Artificial Intelligence Fall 2005 Handout #7 Perceptrons

COMP 551 Applied Machine Learning Lecture 14: Neural Networks

Master Recherche IAC Apprentissage Statistique, Optimisation & Applications

Pattern Recognition Prof. P. S. Sastry Department of Electronics and Communication Engineering Indian Institute of Science, Bangalore

Transcription:

Computational Intelligence Winter Term 2009/10 Prof. Dr. Günter Rudolph Lehrstuhl für Algorithm Engineering (LS 11) Fakultät für Informatik TU Dortmund

Plan for Today Organization (Lectures / Tutorials) Overview CI Introduction to ANN McCulloch Pitts Neuron (MCP) Minsky / Papert Perceptron (MPP) 2

Organizational Issues Who are you? either studying Automation and Robotics (Master of Science) Module Optimization or studying Informatik - BA-Modul Einführung in die Computational Intelligence - Hauptdiplom-Wahlvorlesung (SPG 6 & 7) 3

Organizational Issues Who am I? Günter Rudolph Fakultät für Informatik, LS 11 Guenter.Rudolph@tu-dortmund.de OH-14, R. 232 best way to contact me if you want to see me office hours: Tuesday, 10:30 11:30am and by appointment 4

Organizational Issues Lectures Wednesday 10:15-11:45 OH-14, R. E23 Tutorials Wednesday 16:15-17:00 OH-14, R. 304 group 1 Thursday 16:15-17:00 OH-14, R. 304 group 2 Tutor Nicola Beume, LS11 Information http://ls11 www.cs.unidortmund.de/people/rudolph/ teaching/lectures/ci/ws2009 10/lecture.jsp Slides Literature see web see web 5

Prerequisites Knowledge about mathematics, programming, logic is helpful. But what if something is unknown to me? covered in the lecture pointers to literature... and don t hesitate to ask! 6

Overview Computational Intelligence What is CI? umbrella term for computational methods inspired by nature artifical neural networks evolutionary algorithms fuzzy systems swarm intelligence artificial immune systems growth processes in trees... backbone new developments 7

Overview Computational Intelligence term computational intelligence coined by John Bezdek (FL, USA) originally intended as a demarcation line establish border between artificial and computational intelligence nowadays: blurring border our goals: 1. know what CI methods are good for! 2. know when refrain from CI methods! 3. knowwhytheyworkat all! 4. know how to apply and adjust CI methods to your problem! 8

Biological Prototype Neuron - Information gathering (D) - Information processing (C) - Information propagation (A / S) human being: 10 12 neurons electricity in mv range speed: 120 m / s cell body (C) axon (A) nucleus dendrite (D) synapse (S) 9

Abstraction dendrites nucleus / cell body axon synapse signal input signal processing signal output 10

Model x 2 funktion f f(, x 2,, x n ) x n McCulloch-Pitts-Neuron 1943: x i { 0, 1 } =: B f: B n B 11

1943: Warren McCulloch / Walter Pitts description of neurological networks modell: McCulloch-Pitts-Neuron (MCP) basic idea: - neuron is either active or inactive - skills result from connecting neurons considered static networks (i.e. connections had been constructed and not learnt) 12

McCulloch-Pitts-Neuron n binary input signals,, x n threshold > 0 boolean OR boolean AND can be realized: x 2... 1 x 2... n x n = 1 x n = n 13

McCulloch-Pitts-Neuron n binary input signals,, x n threshold > 0 in addition: m binary inhibitory signals y 1,, y m NOT 0 y 1 if at least one y j = 1, then output = 0 otherwise: - sum of inputs threshold, then output = 1 else output = 0 14

Analogons Neurons Synapse Topology simple MISO processors (with parameters: e.g. threshold) connection between neurons (with parameters: synaptic weight) interconnection structure of net Propagation Training / Learning working phase of ANN processes input to output adaptation of ANN to certain data 15

Assumption: inputs also available in inverted form, i.e. inverted inputs. x 2 + x 2 Theorem: Every logical function F: B n B can be simulated with a two-layered McCulloch/Pitts net. Example: x 2 x 3 x 2 x 3 x 4 3 3 2 1 16

Proof: (by construction) Every boolean function F can be transformed in disjunctive normal form 2 layers (AND - OR) 1. Every clause gets a decoding neuron with = n output = 1 only if clause satisfied (AND gate) 2. All outputs of decoding neurons are inputs of a neuron with = 1 (OR gate) q.e.d. 17

Generalization: inputs with weights x 2 0,2 0,4 0,3 0,7 fires 1 if 0,2 + 0,4 x 2 + 0,3 x 3 0,7 2 + 4 x 2 + 3 x 3 7 10 x 3 duplicate inputs! x 2 7 x 3 equivalent! 18

Theorem: Weighted and unweighted MCP-nets are equivalent for weights Q +. Proof: Let N Multiplication with yields inequality with coefficients in N Duplicate input x i, such that we get a i b 1 b 2 b i-1 b i+1 b n inputs. Threshold = a 0 b 1 b n Set all weights to 1. q.e.d. 19

Conclusion for MCP nets + feed-forward: able to compute any Boolean function + recursive: able to simulate DFA very similar to conventional logical circuits difficult to construct no good learning algorithm available 20

Perceptron (Rosenblatt 1958) complex model reduced by Minsky & Papert to what is necessary Minsky-Papert perceptron (MPP), 1969 What can a single MPP do? J 1 isolation of x 2 yields: J 1 N 0 N 0 Example: 1 0 N J 0 1 separating line separates R 2 in 2 classes 21

= 0 = 1 AND OR NAND NOR 1 0 0 1 1 0 XOR 0 1? x 2 xor 0 0 0 0 1 1 1 0 1 1 1 0 w 1 + w 2 x 2 0 < w 2 w 1 w 1 + w 2 < w 1, w 2 > 0 w 1 + w 2 2 contradiction! 22

1969: Marvin Minsky / Seymor Papert book Perceptrons analysis math. properties of perceptrons disillusioning result: perceptions fail to solve a number of trivial problems! - XOR-Problem - Parity-Problem - Connectivity-Problem conclusion : All artificial neurons have this kind of weakness! research in this field is a scientific dead end! consequence: research funding for ANN cut down extremely (~ 15 years) 23

how to leave the dead end : 1. Multilayer Perceptrons: x 2 2 x 2 2 1 realizes XOR 2. Nonlinear separating functions: XOR g(, x 2 ) = 2 + 2x 2 4 x 2-1 with = 0 1 0 0 1 g(0,0) = 1 g(0,1) = +1 g(1,0) = +1 g(1,1) = 1 24

How to obtain weights w i and threshold? as yet: by construction example: NAND-gate x 2 NAND 0 0 1 1 0 1 0 1 1 1 1 0 0 w 2 w 1 w 1 + w 2 < requires solution of a system of linear inequalities ( P) (e.g.: w 1 = w 2 = -2, = -3) now: by learning / training 25

Perceptron Learning Assumption: test examples with correct I/O behavior available Principle: (1) choose initial weights in arbitrary manner (2) fed in test pattern (3) if output of perceptron wrong, then change weights (4) goto (2) until correct output for al test paterns graphically: translation and rotation of separating lines 26

Perceptron Learning P: set of positive examples N: set of negative examples 1. choose w 0 at random, t = 0 2. choose arbitrary x P N 3. if x P and w t x > 0 then goto 2 if x N and w t x 0 then goto 2 4. if x P and w t x 0 then w t+1 = w t + x; t++; goto 2 5. if x Nand w t x > 0 then w t+1 = w t x; t++; goto 2 I/O correct! let w x 0, should be > 0! (w+x) x = w x + x x > w x let w x > 0, should be 0! (w x) x = w x x x < w x 6. stop? If I/O correct for all examples! remark: algorithm converges, is finite, worst case: exponential runtime 27

Example threshold as a weight: w = (, w 1, w 2 ) 1 x 2 - w 1 w 2 0 suppose initial vector of weights is w (0) = (1, -1, 1) 28

We know what a single MPP can do. What can be achieved with many MPPs? Single MPP separates plane in two half planes Many MPPs in 2 layers can identify convex sets A B 1. How? 2 layers! 2. Convex? a,b X: a + (1- ) b X for (0,1) 29

Single MPP separates plane in two half planes Many MPPs in 2 layers can identify convex sets Many MPPs in 3 layers can identify arbitrary sets Many MPPs in > 3 layers not really necessary! arbitrary sets: 1. partitioning of nonconvex set in several convex sets 2. two-layered subnet for each convex set 3. feed outputs of two-layered subnets in OR gate (third layer) 30