Multi-Dimensional Neural Networks: Unified Theory

Similar documents
COMP9444 Neural Networks and Deep Learning 2. Perceptrons. COMP9444 c Alan Blair, 2017

Introduction to Neural Networks

Artificial Neural Networks. Part 2

Neural networks. Chapter 19, Sections 1 5 1

Reification of Boolean Logic

Artificial Neural Networks

CE213 Artificial Intelligence Lecture 13

Neural Networks. Chapter 18, Section 7. TB Artificial Intelligence. Slides from AIMA 1/ 21

ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD

Lecture 7 Artificial neural networks: Supervised learning

Introduction To Artificial Neural Networks

Artificial Neural Networks. Historical description

Neural networks. Chapter 20, Section 5 1

Computational Intelligence Winter Term 2009/10

Plan. Perceptron Linear discriminant. Associative memories Hopfield networks Chaotic networks. Multilayer perceptron Backpropagation

Optimization of Quadratic Forms: NP Hard Problems : Neural Networks

Computational Intelligence

Lab 5: 16 th April Exercises on Neural Networks

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Neural networks. Chapter 20. Chapter 20 1

Last updated: Oct 22, 2012 LINEAR CLASSIFIERS. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Lecture 4: Perceptrons and Multilayer Perceptrons

Introduction to Artificial Neural Networks

Lecture 4: Feed Forward Neural Networks

EEE 241: Linear Systems

Machine Learning. Neural Networks

Simple Neural Nets For Pattern Classification

Computational Intelligence

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

Machine Learning. Neural Networks. Le Song. CSE6740/CS7641/ISYE6740, Fall Lecture 7, September 11, 2012 Based on slides from Eric Xing, CMU

AN INTRODUCTION TO NEURAL NETWORKS. Scott Kuindersma November 12, 2009

Last update: October 26, Neural networks. CMSC 421: Section Dana Nau

Dynamics of Structured Complex Recurrent Hopfield Networks

CSC Neural Networks. Perceptron Learning Rule

Networks of McCulloch-Pitts Neurons

Neural Networks (Part 1) Goals for the lecture

CS:4420 Artificial Intelligence

Unit 8: Introduction to neural networks. Perceptrons

Simple Neural Nets for Pattern Classification: McCulloch-Pitts Threshold Logic CS 5870

Artificial Neural Network

Neural Networks Lecturer: J. Matas Authors: J. Matas, B. Flach, O. Drbohlav

Neural Networks. Fundamentals of Neural Networks : Architectures, Algorithms and Applications. L, Fausett, 1994

Artificial Neural Networks Examination, June 2004

Learning and Memory in Neural Networks

CHALMERS, GÖTEBORGS UNIVERSITET. EXAM for ARTIFICIAL NEURAL NETWORKS. COURSE CODES: FFR 135, FIM 720 GU, PhD

CS 4700: Foundations of Artificial Intelligence

Artificial Neural Networks Examination, March 2004

ARTIFICIAL INTELLIGENCE. Artificial Neural Networks

Multilayer Perceptron Tutorial

Grundlagen der Künstlichen Intelligenz

PMR5406 Redes Neurais e Lógica Fuzzy Aula 3 Single Layer Percetron

18.6 Regression and Classification with Linear Models

Hopfield-Amari Neural Network:

Artificial Neural Networks. Introduction to Computational Neuroscience Tambet Matiisen

CS 256: Neural Computation Lecture Notes

Neural Networks: Introduction

Artificial Neural Networks Examination, June 2005

Single layer NN. Neuron Model

Artificial Neural Networks. Q550: Models in Cognitive Science Lecture 5

Course 395: Machine Learning - Lectures

Artificial Intelligence

COMP 551 Applied Machine Learning Lecture 14: Neural Networks

Linear & nonlinear classifiers

Neural Networks: Basics. Darrell Whitley Colorado State University

Chapter 9: The Perceptron

Neural Networks biological neuron artificial neuron 1

Neural Networks and Deep Learning

Artificial Neural Networks

Artificial Neural Networks" and Nonparametric Methods" CMPSCI 383 Nov 17, 2011!

Artificial neural networks

Introduction Biologically Motivated Crude Model Backpropagation

Perceptron. (c) Marcin Sydow. Summary. Perceptron

y k = (a)synaptic f(x j ) link linear i/p o/p relation (b) Activation link linear i/p o/p relation

Engineering Part IIB: Module 4F10 Statistical Pattern Processing Lecture 5: Single Layer Perceptrons & Estimating Linear Classifiers

22c145-Fall 01: Neural Networks. Neural Networks. Readings: Chapter 19 of Russell & Norvig. Cesare Tinelli 1

17 Neural Networks NEURAL NETWORKS. x XOR 1. x Jonathan Richard Shewchuk

Sections 18.6 and 18.7 Artificial Neural Networks

Neural Networks for Machine Learning. Lecture 2a An overview of the main types of neural network architecture

CMSC 421: Neural Computation. Applications of Neural Networks

Computational Intelligence Winter Term 2017/18

Neural Networks. Nicholas Ruozzi University of Texas at Dallas

Artificial Neural Networks The Introduction

Artificial Neural Networks

Convolutional Associative Memory: FIR Filter Model of Synapse

CSCI 315: Artificial Intelligence through Deep Learning

CSE20: Discrete Mathematics for Computer Science. Lecture Unit 2: Boolan Functions, Logic Circuits, and Implication

Computational Intelligence

ARTIFICIAL NEURAL NETWORKS گروه مطالعاتي 17 بهار 92

On the Dynamics of a Recurrent Hopfield Network

XOR - XNOR Gates. The graphic symbol and truth table of XOR gate is shown in the figure.

Are Rosenblatt multilayer perceptrons more powerfull than sigmoidal multilayer perceptrons? From a counter example to a general result

Sections 18.6 and 18.7 Artificial Neural Networks

Feedforward Neural Nets and Backpropagation

CS 4700: Foundations of Artificial Intelligence

SPSS, University of Texas at Arlington. Topics in Machine Learning-EE 5359 Neural Networks

AI Programming CS F-20 Neural Networks

Linear & nonlinear classifiers

Simulating Neural Networks. Lawrence Ward P465A

CS 590D Lecture Notes

Transcription:

Multi-Dimensional Neural Networks: Unified Theory Garimella Ramamurthy Associate Professor IIIT-Hyderebad India Slide 1

Important Publication Book based on my MASTERPIECE Title: Multi-Dimensional Neural Networks: Unified Theory Publisher: New Age International Publishers Slide 2

Outline of the Talk 1. One-dimensional neural networks McCulloch Pitts Perceptron (single layer) XOR problem Multi-layer Perceptron Hopfield neural network 2. One-dimensional neural networks Boolean logic theory 3. One-dimensional neural networks Coding theory 4. One-dimensional neural networks Control theory Slide 3

1. Multi-dimensional logic theory Neural networks 2. Multi-dimensional coding theory Neural networks 3. Tensor state space representation 4. Multi-dimensional control theory Neural networks 5. Unified theory in multi-dimensions 6. Implications Slide 4

1. One-Dimensional Neural Networks Goal: To synthesize units which mimic the functions of biological neurons McCulloch Pitts Neuron: x 2 x 1. w 1 w 2 Artificial neuron y x N-1 x N w N-1 w N y = Sign N j = 1 x w j j Limitation: - Doesn t mimic all the functions of a biological neuron. - No training. Slide 5

Perceptron (Rosenblatt) Incorporation of training by variation of weights, i.e., w j (n+1) = w j (n) + η x j ( t o ); t target output o.actual output Convergence of weights using the learning law. x 1 x 2. w N1 w 11 w 21 w 22 w 12 neuron 1 i.e., the input pattern space is divided Into several regions using hyperplanes. x N w N2 neuron 2 Slide 6

Minsky: XOR problem x 2 XOR gate (0,1) (1,1) (0,0) (1,0) x 1 i.e., it is concluded that XOR gate cannot be synthesized using a single layer Perceptron. Slide 7

Multi-Layer Perceptron:......... Input layer Hidden layer Output layer Hornik Stinchcombe White Theorem. i.e., even nonlinearly separable patterns can be classified using a multi-layer Perceptron, i.e., XOR problem disappears. Functional Link Networks: patterns not linearly separable in a lower dimensional space can be linearly separable in a higher dimensional space. Slide 8

Hopfield/Amari Neural Network: G(V,E ) : Fully connected undirected graph - Let there be N nodes/neurons ( V = N ) N V ( n + 1) = Sign W V ( n ) T i ij j i j = 1 - Modes of operation: - serial mode - parallel mode - Convergence theorem: - In the serial mode network always converges starting in an initial state. - In parallel mode convergence or cycle of length 2 appears. - Energy function is non-decreasing. Slide 9

Control, Communication and Computation Theories Control Theory : Move a system from one point in state space to another point such that certain objective function is optimized Communication Theory: Convey a message from one point in space to another point reliably. Slide 10

Control, Computation Theories: Neural Nets Computation Theory: Process a set of input symbols and produce a set of output symbols based on some information processing operation Question: Are the above theories related to the THEORY of NEURAL NETWORKS Slide 11

2. One-Dimensional Neural Networks: Boolean logic theory Logic Gates: AND, OR, NOT, NAND, NOR, XOR (2 inputs or multiple inputs). Relationship between neural networks and logic gates: (Chakradhar, et al.). Given a logic gate, there exists a neural network such that the inputs are mapped to stable states, (which constitute the logic gate outputs). i.e., A Hopfield network can be utilized to synthesize a logic gate. Slide 12

1. One-Dimensional Neural Networks: Coding theory Given a Hopfield Neural Network (optimizing quadratic energy function), there exists a graph-theoretic code (encoder) such that the input information vector is mapped to a codeword (and vice versa). Give a generalized neural network (optimizing an energy function higher than a quadratic form), there exists a linear code (encoder) such that every stable state (mapping an input information vector) corresponds to a codeword (and vice versa). Also maximum likelihood decoding corresponds to determining the global optimum of energy function. Slide 13

4. One-Dimensional Neural Networks: Control theory Problem Statement: Given a linear time varying system, determine a finite sequence of input values (bounded in magnitude by unity). Such that the total output energy over a finite horizon is as maximum as possible. Solution: (Using Pontriyagin s maximum principle). The optimal input vector constitutes the stable state of a Hopfield neural network, i.e., u = Sign( Ru ) where R Energy Density Matrix of Linear System. Slide 14

5. Multi-Dimensional Logic Theory: Neural Networks Q: How can logic gate definition be extended to a two-dimensional array of 0 s and 1 s? Some possibilities in the case of a matrix of 0 s and 1 s. Such an approach is heuristic and the logic gate definition is non-unique. Difficulties are compounded for a multidimensional logic gate definition Generalization of one-dimensional definition. Slide 15

m-d array of inputs Multi-dimensional Logic gate m-d logic gate outputs m-d array of inputs Multi-dimensional Neural network m-d neural network stable state Energy Function: Quadratic or higher degree form. Slide 16

6. Multi-Dimensional Coding Theory: Neural Networks Multi-dimensional Hopfield neural network. Maximization of energy function (quadratic form) defined on the m-d hypercube. Encoder of a multi-dimensional graph theoretic code. (Application: Multi-dimensional associative memory.) Multi-dimensional generalized neural network. Maximization of higher degree energy function on the multi-dimensional bounded lattice. Encoder of a multidimensional linear/non-linear code m-d Stable states codewords. Slide 17

7. Tensor State Space Representation: Multi-dimensional system theory One-dimensional linear systems. State space representation (Kalman) Earlier efforts in 2-dimensional system theory: (applications in image processing, etc.). No notion of causality quarter plane and half plane causality. Slide 18

Main Contribution: Tensor state space representation A X ( t ) i 1,i 2,,in; j 1, j 2,, jn j 1, j 2,, jn + B U ( t ) = X ( t ) i 1,,in; j 1, jn j 1,, jn i 1, in Y ( t ) = C X ( t ) j 1, jn i 1,,in; j 1,, jn j 1,, jn + D U ( t ) j 1,, jn j 1,, jn Slide 19

8. Multi-Dimensional Control Theory: Neural networks Problem formulation: Given a multi-dimensional linear system with TSSR, with input tensors subjected to a bounded amplitude constraint, maximize the TOTAL output energy over a FINITE horizon. Solution: The optimal control tensors constitute the stable states of a multi-dimensional Hopfield neural network. concepts such as controllability, observability, stability are generalized from one-dimension to multi-dimensions with TSSR. Slide 20

9. Unified Theory of Control: Communication and computation Mathematical cybernetics: Optimal control tensors, optimal codeword tensors, optimal switching function tensors constitute the stable states of generalized multi-dimensional neural networks. optimal control tensors stable states of m-d neural networks optimal m-d logic gate output tensors optimal codeword tensors Slide 21

10. Implications: Applications: 3-d/m-d associative memory. Powerful robots. Theoretical implications: Relating work on topological groups and maximum principle. Functional analysis. Infinite dimensional generalization. Slide 22