Associative Memories

Similar documents
CHAPTER III Neural Networks as Associative Memory

Cyclic Codes BCH Codes

Unsupervised Learning

Lecture 3: Shannon s Theorem

Internet Engineering. Jacek Mazurkiewicz, PhD Softcomputing. Part 3: Recurrent Artificial Neural Networks Self-Organising Artificial Neural Networks

Example: Suppose we want to build a classifier that recognizes WebPages of graduate students.

MARKOV CHAIN AND HIDDEN MARKOV MODEL

Hopfield networks and Boltzmann machines. Geoffrey Hinton et al. Presented by Tambet Matiisen

Hopfield Training Rules 1 N

VQ widely used in coding speech, image, and video

Numerical integration in more dimensions part 2. Remo Minero

HOPFIELD NETWORKS 9.1 INTRODUCTION

Note 2. Ling fong Li. 1 Klein Gordon Equation Probablity interpretation Solutions to Klein-Gordon Equation... 2

Research on Complex Networks Control Based on Fuzzy Integral Sliding Theory

Instance-Based Learning (a.k.a. memory-based learning) Part I: Nearest Neighbor Classification

CSC321 Tutorial 9: Review of Boltzmann machines and simulated annealing

find (x): given element x, return the canonical element of the set containing x;

Physics 2A Chapter 3 HW Solutions

Problem Set 9 Solutions

Neural network-based athletics performance prediction optimization model applied research

APPENDIX A Some Linear Algebra

Boundary Value Problems. Lecture Objectives. Ch. 27

Chapter 7 Channel Capacity and Coding

Calculation of time complexity (3%)

More metrics on cartesian products

U.C. Berkeley CS294: Beyond Worst-Case Analysis Luca Trevisan September 5, 2017

NP-Completeness : Proofs

The equation of motion of a dynamical system is given by a set of differential equations. That is (1)

Inner Product. Euclidean Space. Orthonormal Basis. Orthogonal

ON AUTOMATIC CONTINUITY OF DERIVATIONS FOR BANACH ALGEBRAS WITH INVOLUTION

College of Computer & Information Science Fall 2009 Northeastern University 20 October 2009

Which Separator? Spring 1

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Turing Machines (intro)

CS 3710: Visual Recognition Classification and Detection. Adriana Kovashka Department of Computer Science January 13, 2015

Supervised Learning. Neural Networks and Back-Propagation Learning. Credit Assignment Problem. Feedforward Network. Adaptive System.

Supplementary Material: Learning Structured Weight Uncertainty in Bayesian Neural Networks

CHALMERS, GÖTEBORGS UNIVERSITET. SOLUTIONS to RE-EXAM for ARTIFICIAL NEURAL NETWORKS. COURSE CODES: FFR 135, FIM 720 GU, PhD

Solution of Linear System of Equations and Matrix Inversion Gauss Seidel Iteration Method

10-701/ Machine Learning, Fall 2005 Homework 3

Kernel Methods and SVMs Extension

Discussion 11 Summary 11/20/2018

Optimal information storage in noisy synapses under resource constraints

1 Convex Optimization

Chapter Newton s Method

CS407 Neural Computation

Statistical Inference. 2.3 Summary Statistics Measures of Center and Spread. parameters ( population characteristics )

we have E Y x t ( ( xl)) 1 ( xl), e a in I( Λ ) are as follows:

Supporting Information

Dynamic Systems on Graphs

CS 331 DESIGN AND ANALYSIS OF ALGORITHMS DYNAMIC PROGRAMMING. Dr. Daisy Tang

Common loop optimizations. Example to improve locality. Why Dependence Analysis. Data Dependence in Loops. Goal is to find best schedule:

[WAVES] 1. Waves and wave forces. Definition of waves

1 Derivation of Rate Equations from Single-Cell Conductance (Hodgkin-Huxley-like) Equations

Admin NEURAL NETWORKS. Perceptron learning algorithm. Our Nervous System 10/25/16. Assignment 7. Class 11/22. Schedule for the rest of the semester

Chapter 7 Channel Capacity and Coding

Foundations of Arithmetic

THE ARIMOTO-BLAHUT ALGORITHM FOR COMPUTATION OF CHANNEL CAPACITY. William A. Pearlman. References: S. Arimoto - IEEE Trans. Inform. Thy., Jan.

8 Derivation of Network Rate Equations from Single- Cell Conductance Equations

ECE 534: Elements of Information Theory. Solutions to Midterm Exam (Spring 2006)

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification

CSE4210 Architecture and Hardware for DSP

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems

MACHINE APPLIED MACHINE LEARNING LEARNING. Gaussian Mixture Regression

SCALARS AND VECTORS All physical quantities in engineering mechanics are measured using either scalars or vectors.

Week3, Chapter 4. Position and Displacement. Motion in Two Dimensions. Instantaneous Velocity. Average Velocity

arxiv: v1 [cs.lg] 17 Jan 2019

C/CS/Phy191 Problem Set 3 Solutions Out: Oct 1, 2008., where ( 00. ), so the overall state of the system is ) ( ( ( ( 00 ± 11 ), Φ ± = 1

Chapter 6 Hidden Markov Models. Chaochun Wei Spring 2018

9 Derivation of Rate Equations from Single-Cell Conductance (Hodgkin-Huxley-like) Equations

Generalized Linear Methods

LECTURE NOTES. Artifical Neural Networks. B. MEHLIG (course home page)

Single Variable Optimization

Analysis of Bipartite Graph Codes on the Binary Erasure Channel

Support Vector Machines

First Law: A body at rest remains at rest, a body in motion continues to move at constant velocity, unless acted upon by an external force.

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X

PES 1120 Spring 2014, Spendier Lecture 6/Page 1

Finite Difference Method

Chapter 9: Statistical Inference and the Relationship between Two Variables

The Cortex. Networks. Laminar Structure of Cortex. Chapter 3, O Reilly & Munakata.

Chapter 8. Potential Energy and Conservation of Energy

MODELING TRAFFIC LIGHTS IN INTERSECTION USING PETRI NETS

8 Derivation of Network Rate Equations from Single- Cell Conductance Equations

EGR 544 Communication Theory

Delay tomography for large scale networks

A finite difference method for heat equation in the unbounded domain

Image Classification Using EM And JE algorithms

Here is the rationale: If X and y have a strong positive relationship to one another, then ( x x) will tend to be positive when ( y y)

Mean Field / Variational Approximations

Pulse Coded Modulation

The Application of BP Neural Network principal component analysis in the Forecasting the Road Traffic Accident

Lecture 12: Discrete Laplacian

Multilayer Perceptrons and Backpropagation. Perceptrons. Recap: Perceptrons. Informatics 1 CG: Lecture 6. Mirella Lapata

LECTURE 21 Mohr s Method for Calculation of General Displacements. 1 The Reciprocal Theorem

2 Laminar Structure of Cortex. 4 Area Structure of Cortex

A A Non-Constructible Equilibrium 1

Physics 5153 Classical Mechanics. Principle of Virtual Work-1

PHYS 705: Classical Mechanics. Newtonian Mechanics

Logic Learning in Hopfield Networks

Transcription:

Assocatve Memores We consder now modes for unsupervsed earnng probems, caed auto-assocaton probems. Assocaton s the task of mappng patterns to patterns. In an assocatve memory the stmuus of an ncompete or corrupted pattern eads to the response of a stored pattern that corresponds n some manner to the nput pattern. A neura network mode most commony used for (auto-) assocaton probems s the Hopfed network. 5

Eampe Corrupted mage Intermedate state of Hopfed net Output States Bt maps. Attractors Prototype patterns. Input An arbtrary pattern (e.g. pcture wth nose). Output The best prototype for that pattern. 5 2

FIELD ETWORKS The Hopfed network mpements a so-caed content addressabe memory. A coecton of patterns caed fundamenta memores s stored n the by means of weghts. Each neuron represents a component of the nput. The weght of the nk between two neurons measures the correaton between the two correspondng components over the fundamenta memores. If the weght s hgh then the correspondng components are often equa n the fundamenta memores. 5 3

ARCHITECTURE: recurrent z - Mutpe-oop feedback system wth no sef-feedback z - z - unt-deay operator 5 4

Hopfed dscrete Input vectors vaues are n {-,} (or {0,}). The number of neurons s equa to the nput dmenson. Every neuron has a nk from every other neuron (recurrent archtecture) ecept tsef (no sef-feedback). The neuron state at tme n s ts output vaue. The network state at tme n s the vector of neurons states. The actvaton functon used to update a neuron state s the sgn functon but f the nput of the actvaton functon s 0 then the new output (state) of the neuron s equa to the od one. Weghts are symmetrc: w = w 5 5

otaton : nput dmenson. M: number of fundamenta memores. f -th component of the µ µ fundamenta memory. (n) State of neuron at tme n. 5 6

Weghts computaton. Storage. Let f, f 2,, f M denote a known set of - dmensona fundamenta memores. The synaptc weghts of the network are: w M M = µ = 0 f µ, f µ, =, =... where w s the weght from neuron to neuron. The eements of the vectors f µ are n {-,+}. Once computed, the synaptc weghts are fed. 5 7

Eecuton 2. Intasaton. Let probe denote an nput vector (probe) presented to the network. The agorthm s ntased by settng: where (0) s the state of neuron at tme n = 0, and probe, s the -th eement of the probe vector probe. 3. Iteraton Unt Convergence. Update the eements of network state vector (n) asynchronousy (.e. randomy and one at the tme) accordng to the rue ( n + ) = sgn w = ( n) Repeat the teraton unt the state vector remans unchanged. fed 4. Outputtng. Let denote the fed pont (or stabe state, that s such that (n+)=(n)) computed at the end of step 3. The resutng output y of the network s: y = ( 0) = probe, =,..., fed =, 2,..., 5 8

Eampe weght - - + + (-,, -) (-,, ) (,, -) (,, ) - 2 3 - neuron (-, -, ) (, -, ) stabe states - - (-, -, -) (, -, -) - - - - - - - - - attracton basn attracton basn 2-5 9

Eampe 2 Separaton of patterns usng the two fundamenta memores (- - -) ( ): Fnd weghts to obtan the foowng behavor: - - - - - - - - - - - - 2 3 w =? 5 0

COVERGECE Every stabe state s at an energy mnmum. (A state s stabe f (n+)=(n)) What s energy? Energy s a functon (Lyapunov functon): E: States such that every frng (change of ts output vaue) of a neuron decreases the vaue of E: mpes E( ) < E() E( ) = = 2 2 T, W w 5

COVERGECE A stored pattern estabshes a correaton between pars of neurons: neurons tend to have the same state or opposte state accordng to ther vaues n the pattern. If w s arge, ths epresses an epectaton that neurons and are postvey correated smary. If t s sma (very negatve) ths ndcates a negatve correaton. w w thus be arge for a state, whch s a stored, pattern (snce w w be postve f > 0 and negatve f < 0). The negatve of the sum w thus be sma. 5 2

Energy Decreases Cam: Frng a neuron decreases the energy E. Proof: Let be the neuron that fres: Ether:. goes from - to whch mpes = w > 0 (above threshod) 2. goes from to - whch mpes = w < 0 (beow threshod) Thus n both cases: ( ' ) w > 0 vaue after = 5 3

5 4 Proof ow we try to factor from the energy functon E: + = = = = = = = = = = = = = = = = = w w w w w w w 2 2 2 2 2 2 2 w E pued out the = case can be added snce w = 0 Term ndependent of

5 5 Proof How does E change wth changng? = + = = = = = = E E E E w ) ' ( w w ' ' E after change vaue of ' E before change vaue of we showed n prevous sde that ths quantty s > 0 (wthout - sgn) So E -E < 0 aways.

Convergence resut We have shown that the energy E decreases wth each neuron frng. The overa number of states s fnte {+, -}. Then The energy cannot decrease forever. Frng cannot contnue forever. 5 6

Computer eperment We ustrate the behavor of the dscrete Hopfed network as a content addressabe memory. n = 20 neurons ( n 2 - n = 4280 weghts). The network s traned to retreve 8 back and whte patterns. Each pattern contans 20 pes. The nputs of the net assume vaue + for back pes and - for whte pes. Retreva : the fundamenta memores are presented to the network to test ts abty to recover them correcty from the nformaton stored n the weghts. In each case, the desred pattern was produced by the network after teraton. 5 7

Computer eperment These patterns are used as fundamenta memores to create the weght matr. 5 8

Computer eperment Retreva 2: to test error-correctng capabty of the network. A pattern of nterest s dstorted by randomy and ndependenty reversng each pe of the pattern wth a probabty of 0.25. The dstorted pattern s used as a probe for the network. The average number of teratons needed to reca, averaged over the eght patterns, s about 3. The net behaves as epected. 5 9

Computer eperment 5 20

Computer eperment 5 2

Spurous attractor states Probem of ncorrect retreva. For eampe the network wth nput corrupted pattern 2 converges to 6. Spurous attractor states: the net sde shows 08 spurous attractors found n 43097 tests of randomy seected dgts corrupted wth the probabty of fppng a bt at 0.25. 5 22

Spurous attractor states Combnaton of dgt, dgt 4 and dgt 9 Loca mnma of E not correated wth any of the fundamenta memores 5 23

Storage Capacty Storage capacty: the quantty of nformaton that can be stored n a network n such a way that t can be retreved correcty. Defnton of storage capacty: C = number of fundamenta patterns number of neurons n the network C = number of fundamenta patterns number of weghtsn the network 5 24

Storage Capacty: bounds Theorem: The mamum storage capacty of a dscrete Hopfed s bounded above by C M = 4 n That s, f β s the probabty that the -th bt of the m-th fundamenta memory s correcty retreved for a =,.., and m =,,M, then m β = whenever M < / (4 n ) 5 25

Hopfed for Optmzaton Optmzaton probems (ke Traveng Saesman) can be encoded nto Hopfed etworks Obectve functon corresponds to the energy of the network Good soutons are stabe states of the network 5 26

TSP Traveng Saesman Probem (TSP): Gven ctes wth dstances d. What s the shortest tour? 5 27

Encodng Construct a Hopfed network wth 2 nodes. Semantcs: na = ff town s vsted at step a Constrants: na =, a na =, a The town dstances are encoded by weghts,.e. w = ab d 5 28

Hopfed etwork for TSP pace cty 2 3 4 A 0 0 0 B 0 0 0 C 0 0 0 D 0 0 0 Confguraton not aowed Confguraton aowed 5 29

Energy and Weghts =,, a, b odes wthn each row ( = ) connected wth weghts w ab E = γ ab odes wthn each coumn ( a=b ) connected wth weghts w ab 2 = γ Each node s connected to nodes n coumns eft and rght wth weght w n a n b w = d ab 5 30