Storage Capacity of Letter Recognition in Hopfield Networks

Similar documents
SCRAM: Statistically Converging Recurrent Associative Memory

arxiv: v2 [nlin.ao] 19 May 2015

Hopfield Neural Network and Associative Memory. Typical Myelinated Vertebrate Motoneuron (Wikipedia) Topic 3 Polymers and Neurons Lecture 5

Artificial Intelligence Hopfield Networks

Consider the way we are able to retrieve a pattern from a partial key as in Figure 10 1.

Terminal attractor optical associative memory with adaptive control parameter

Hopfield Networks. (Excerpt from a Basic Course at IK 2008) Herbert Jaeger. Jacobs University Bremen

7 Recurrent Networks of Threshold (Binary) Neurons: Basis for Associative Memory

Hopfield Neural Network

Properties of Associative Memory Model with the β-th-order Synaptic Decay

Storkey Learning Rules for Hopfield Networks

7 Rate-Based Recurrent Networks of Threshold Neurons: Basis for Associative Memory

T sg. α c (0)= T=1/β. α c (T ) α=p/n

Learning and Memory in Neural Networks

In biological terms, memory refers to the ability of neural systems to store activity patterns and later recall them when required.

THIRD-ORDER HOPFIELD NETWORKS: EXTENSIVE CALCULATIONS AND SIMULATIONS

Hopfield Network for Associative Memory

Storage capacity of hierarchically coupled associative memories

Using Variable Threshold to Increase Capacity in a Feedback Neural Network

Lecture 14 Population dynamics and associative memory; stable learning

CHAPTER 3. Pattern Association. Neural Networks

Memory capacity of networks with stochastic binary synapses

Associative Memories (I) Hopfield Networks

Memories Associated with Single Neurons and Proximity Matrices

( ) T. Reading. Lecture 22. Definition of Covariance. Imprinting Multiple Patterns. Characteristics of Hopfield Memory

Using a Hopfield Network: A Nuts and Bolts Approach

Application of hopfield network in improvement of fingerprint recognition process Mahmoud Alborzi 1, Abbas Toloie- Eshlaghy 1 and Dena Bazazian 2

Neural Nets and Symbolic Reasoning Hopfield Networks

Computational Intelligence Lecture 6: Associative Memory

A. The Hopfield Network. III. Recurrent Neural Networks. Typical Artificial Neuron. Typical Artificial Neuron. Hopfield Network.

A. The Hopfield Network. III. Recurrent Neural Networks. Typical Artificial Neuron. Typical Artificial Neuron. Hopfield Network.

Neural networks that use three-state neurons

Content-Addressable Memory Associative Memory Lernmatrix Association Heteroassociation Learning Retrieval Reliability of the answer

CSE 5526: Introduction to Neural Networks Hopfield Network for Associative Memory

HOPFIELD neural networks (HNNs) are a class of nonlinear

HOPFIELD MODEL OF NEURAL NETWORK

Week 4: Hopfield Network

arxiv:cond-mat/ v1 [cond-mat.dis-nn] 30 Sep 1999

Neural Networks Lecture 6: Associative Memory II

COMP9444 Neural Networks and Deep Learning 11. Boltzmann Machines. COMP9444 c Alan Blair, 2017

A Cell Assembly Model of Sequential Memory

Neural Networks. Hopfield Nets and Auto Associators Fall 2017

ASYMPTOTICALLY STABLE MULTI-VALUED MANY-TO-MANY ASSOCIATIVE MEMORY NEURAL NETWORK AND ITS APPLICATION IN IMAGE RETRIEVAL

Convolutional Associative Memory: FIR Filter Model of Synapse

Unit III. A Survey of Neural Network Model

Neural Networks. Prof. Dr. Rudolf Kruse. Computational Intelligence Group Faculty for Computer Science

Chapter 10 Associative Memory Networks We are in for a surprise though!

A recurrent model of transformation invariance by association

How can ideas from quantum computing improve or speed up neuromorphic models of computation?

arxiv:cond-mat/ v1 [cond-mat.dis-nn] 18 Nov 1996

Pattern Association or Associative Networks. Jugal Kalita University of Colorado at Colorado Springs

Synaptic plasticity in neuromorphic hardware. Stefano Fusi Columbia University

Generalization in a Hopfield network

Strong Attractors of Hopfield Neural Networks to Model Attachment Types and Behavioural Patterns

2- AUTOASSOCIATIVE NET - The feedforward autoassociative net considered in this section is a special case of the heteroassociative net.

Neural Networks for Machine Learning. Lecture 11a Hopfield Nets

Machine Learning. Neural Networks

Supervised (BPL) verses Hybrid (RBF) Learning. By: Shahed Shahir

Introduction to Neural Networks

Memory capacity of neural networks learning within bounds

ESANN'1999 proceedings - European Symposium on Artificial Neural Networks Bruges (Belgium), April 1999, D-Facto public., ISBN X, pp.

The Generalized Brain-State-in-a-Box (gbsb) Neural Network: Model, Analysis, and Applications

Lecture 4: Feed Forward Neural Networks

(a) (b) (c) Time Time. Time

Associative content-addressable networks with exponentially many robust stable states

Hopfield and Potts-Hopfield Networks

Slide10 Haykin Chapter 14: Neurodynamics (3rd Ed. Chapter 13)

Application of Associative Matrices to Recognize DNA Sequences in Bioinformatics

Effects of Interactive Function Forms in a Self-Organized Critical Model Based on Neural Networks

Effects of refractory periods in the dynamics of a diluted neural network

Artificial Neural Networks Examination, March 2004

Extending the Associative Rule Chaining Architecture for Multiple Arity Rules

Robust exponential memory in Hopfield networks

Neural networks: Unsupervised learning

Hopfield networks. Lluís A. Belanche Soft Computing Research Group

The Variance of Covariance Rules for Associative Matrix Memories and Reinforcement Learning

Neural Network with Ensembles

Autoassociative Memory Retrieval and Spontaneous Activity Bumps in Small-World Networks of Integrate-and-Fire Neurons

Effects of Interactive Function Forms and Refractoryperiod in a Self-Organized Critical Model Based on Neural Networks

Introduction to Artificial Neural Networks

Neural Network Analysis of Russian Parliament Voting Patterns

Logic Learning in Hopfield Networks

Neural Networks. Associative memory 12/30/2015. Associative memories. Associative memories

Information storage capacity of incompletely connected associative memories

On the Hopfield algorithm. Foundations and examples

Hertz, Krogh, Palmer: Introduction to the Theory of Neural Computation. Addison-Wesley Publishing Company (1991). (v ji (1 x i ) + (1 v ji )x i )

Specification and Implementation of Digital Hopfield-Type Associative Memory with On-Chip Training

Statistical Mechanics of Temporal Association in Neural Networks with Delayed Interactions

Neural Network Essentials 1

Rate- and Phase-coded Autoassociative Memory

patterns (\attractors"). For instance, an associative memory may have tocomplete (or correct) an incomplete

AI Programming CS F-20 Neural Networks

Error Correcting Capability of Cellular Automata Based Associative Memory

Lecture 7 Artificial neural networks: Supervised learning

Neural Networks Lecture 2:Single Layer Classifiers

Markov Transitions between Attractor States in a Recurrent Neural Network

Serious limitations of (single-layer) perceptrons: Cannot learn non-linearly separable tasks. Cannot approximate (learn) non-linear functions

W memory based upon ideas from neural nets and

Effect of Connectivity in Associative Memory Models

Dense Associative Memory for Pattern Recognition

Transcription:

Storage Capacity of Letter Recognition in Hopfield Networks Gang Wei (gwei@cs.dal.ca) Zheyuan Yu (zyu@cs.dal.ca) Faculty of Computer Science, Dalhousie University, Halifax, N.S., Canada B3H 1W5 Abstract: In this paper, the Hopfield neural networks model is discussed and implemented for letter recognition. Hebbian and pseudo-inverse learning rules are applied. Comparisons are made between these two rules. The storage capacity, basin of attractions are studied. Keywords: auto-associative memory, Hopfield neural networks, Hebbian, pseudoinverse, learning rule, storage capacity, basin of attraction. 1 INTRODUCTION Associative memory is a dynamical system which has a number of stable states with a domain of attraction around them [1]. If the system starts at any state in the domain, it will converge to the locally stable state, which is called an attractor. In 1982, Hopfield [2] proposed a fully connected neural network model of associative memory in which patterns can be stored by distributed among neurons, and we can retrieve one of the previously presented patterns from an example which is similar to, or a noisy version of it. The Hopfild model consists of N neurons and N two states. It is based on the difference equation s i ( t + 1) = sign[ wis ( t)] 2 synapses. Each neuron can be in one of Where s (t) is the state of neuron i. w is the synaptic strength from neuron to i. The i i appropriate determined synaptic weights can store patterns as fix points. The previously stored patterns can be retrieved from the representation of partial patterns. The dynamical behavior of its neuron states strongly depends on synaptic strength between neurons. The specification of the synaptic weights is conventionally referred to as learning. Hopfield used the Hebbian learning rule [3] to determine weights. Since Hopfield s proposal, many alternative algorithms for learning and associative recalling have been proposed to improve the performance of the Hopfield networks. We will discuss the Hebbian rule and pseudo-inverse rule and apply them to letter recognition. The comparisons are made between these two rules. 2 Different learning rules

Hebbian rule The Hebbian learning rule is given by : v v v 1 1 v v w i = 0 i, and wi = wi ξi ξ N u v Where w is the state of weight matrix after the vth patters have been learnt but before the v (v+1)th pattern has been introduced, ξ is the patterns for the v steps, where N is the number of neurons. The Hebbian rule is local and incremental, but has a low absolute n storage capacity of. This capacity decreases significantly if patterns are correlated. 2ln n Its performance is poor in terms of storage capacity, attraction, and spurious memories [4]. Pseudo-inverse Rule The pseudo-inverse rue is given by m m v 1 v vu u wi = ξ i ( Q 1 ) ξ N v= 1 u= 1 1 n v u Where Q = k = ξ 1 kξ k, and m is the total number of patterns with m<n, N is the N number of neurons. The pseudo-inverse (PI) learning rule (LR), which is also called the proection learning rule [5, 6]. PI LR is accredited for its high retrieval capability. The limit of 0.5N for the associative capacity, obtained by Personnaz et al., Kaner and Sompolinsky [5, 7]. For comparison, Hebbian correlation LR retrieves only up to 14%N prototypes. However, the pseudo-inverse rule is not local or incremental, because it involves the calculation of an inverse. 3 Storage Capacity By using computer simulation, Hopfield suggested that if a small error in recalling is allowed, the maximum number p of the patterns to be stored in a network with N neurons is 0.15N, Later, it was showed theoretically that the storage capacity is p=0.138n by Amit et al. [8] by using replica method. The limitation is attributed to the fact that the network is trapped to the so called spurious local minima. In 1987 McEliece et al. [9] proved that when p < N / 4ln N holds, the Hopfield s model is able to recall all the memorized patterns without error. Other learning schemes have been proposed to escape the local minima and increase the storage capacity. Kohonen et al. [10] proposed a new learning mechanism for non-zero self-connection networks called pseudo-inverse learning rule by extending the Hebbian

learning rule, and enlarged the capacity to p = N. Gardner [11] showed that the ultimate capacity will be p = 2N as the basin size tend to zero. 4 EXPERIMENT All experiments in this paper are carried out on the model of letter patterns. The 26 letter patterns are presented with binary representation (0 or 1) of 156 neurons. The synaptic weight values are explored with Hebbian rule and pseudo-inverse rule. The update scheme is based exactly on the Hopfield model. In the stage of retrieval, 10% noise is applied to letter patterns. Noise is created by randomly inverting pixels of a pattern. We trained the networks with increasing number of letter patterns, which are randomly chosen from 26 letter patterns, and the difference is recorded between the final state of the networks and the corresponding training patterns for 30 repeats. 5 RESULTS AND DISCUSSION 5.1 Storage Capacity Figure 1 Load parameters Vs average Hamming distance between updated states and initial states (Hamming/n) for Hebbian rule. Figure 2 Load parameters Vs average Hamming distance between updated states and initial states (Hamming/n) for pseudo-inverse rule. Figure 1 and 2 show the storage capacity for Hebbian rule and pseudo-inverse rule. The trained patterns are initialized with 10% noise. The average distance is measured by the average Hamming distance between updated states and initial states divided by the number of neurons n=156 for 30 repeat tests. Figure 1 shows that the storage capacity for Hebbian rule, and figure 2 for pseudo-inverse rule. The load parameter is define as N pat pat α = / C, where N is the number of trained pattern, C is the number of nodes in the fully connected networks. According to figure 1 and 2, the storage capacity of Hebbian rule is about 0.012. That means 2 letter patterns can be stored and retrieved without error by the model. And the

storage capacity of pseudo-inverse rule is about 0.064. That means 10 letter patterns can be stored and retrieved without error by the model. Also we notice that the average distance between updated states and initial states for pseudo-inverse rule is much smaller than that for Hebbian rule The storage capacity of our Hopfield networks, for Hebbian rule is 0.012 and for psedoinverse rule is 0.064, are far away from the result in theory which are 0.138 and 1. This can be caused by the following factors [12]. 1) A high correlation of the training patterns. Correlations between the patterns worsen the performance of the network. Pseudo-inverse rule is one of the solutions to lower the correlation of the training patterns by preprocesse the patterns. 2) The capacities of Hopfield networks are related with sparse representation of training patterns. 3) Global inhibition is another factor that can affect storage capacity. 5.2 Basin of Attraction Figure 3 Initial distance (noise level) Vs average Hamming distance between updated states and initial states (Hamming/n) for Hebbian rule. Figure 4 Initial distance (noise level) Vs average Hamming distance between updated states and initial states (Hamming/n) for pseudo-inverse rule Figure 3 and 4 show that the initial distance (noise level) Vs average Hamming distance between updated states and initial states (Hamming/n) for Hebbian rule and pseudo-inverse rule respectively. The distance is measured by Hamming distance between updated states and initial states divided by number of neurons n=156. 2 patterns are chosen randomly from 26 letter patterns and presented to the network for each test, the first trained pattern is presented and the distance is recorded. We can see from figure 3 and 4 that the difference is not significant between the attraction basins for Hebbian rule and that for pseudo-inverse rule at a very low loading.

The pattern completion ability of the associative memory makes the trained patterns point attractors of the network. Figure 3 and 4 demonstrate that, for both Hebbian and pseudoinverse rules, randomly chosen letter pattern converge to desired attractor if the initial distance is less than a certain value d, which is the size of basin of the attraction. Figure 5 Store 2 patterns and retrieve first pattern for pseudo-inverse rule Figure 6 Store 15 letter patterns and retrieve first pattern for pseudo-inverse rule Figure 5 and 6 are similar to figure 3 and 4 but for only store 2 patterns and 15 letter patterns with pseudo-inverse rule respectively. The retrieving of the first letter pattern is recorded. It shows if the networks only store 2 letter patterns, the networks can retrieve the letter pattern without error with the initial distance as long as 0.4. However, if the networks store 15 letter patterns, the longest initial distance decrease to 0.2. That means the size of basin of the attraction will became smaller with more patterns are stored. 6 CONCLUSIONS We have discussed Hopfield model of associative memory and implemented a Hopfield networks for letter recognition. We found that the networks with 156 neurons would store and retrieve without error for only 2 patterns with the Hebbian learning rule in average. The network of this size using pseudo-inverse rule can store an average of 10 patterns and retrieve without error. Also we found that the size of basin of the attraction of the Hopfield networks will became smaller with more stored patterns. REFERENCES [1] J.Komlos, and R. Paturi. Convergence Results in an Associative Memory Model. Neural Networks 1. pp239-250, 1988. [2] J. J. Hopfield. Neural Networks and Physical System with Emergent Collective Computational Abilities. Proceedings of the National Academy of Sciences, USA. 79. pp2554-2558, 1982.

[3] D. O. Hebb. The Organization of Behavior. Wiley, 1949. [4] Storkey A.J. Increasing the capacity of the Hopfield network without sacrificing functionality, ICANN97 p451-456, 1997 [5] L. Personnaz, I. Guyon and G. Dreyfus. Collective computational properties of neural networks: New learning mechanisms, Phys. Rev. A vol 34, pp. 4217-4228, 1986. [6] D.O. Gorodnichy. A way to improve error correction capability of Hopfield associative memory in the case of saturation, HELNET 94-95 International Workshop on Neural Networks Proceedings, vol. I/II, pp. 198-212, VU University Press, Amsterdam, 1996 [7] I. Kanter and H. Sompolinsky. Associative recall of memory without errors, Phys. Rev. A vol 35, pp. 380-392, 1987. [8] D. J. Amit, H. Gutfreund, and H. Sompolinsky. Statistical Mechanics of Neural Networks near Saturation. Annals. of Physics 173, pp30-67, 1985. [9] R. J. McEliced, E. C. Posner, E. R. Rodemick, and S. S. Venkatesh. The capacity of the Hopfield Associative Memory. IEEE Trans. Information Theory IT-33. pp461-482, 1987. [10] T. Kohonen, and M. Ruohonen, Representation of Associated Data by Matrix Operators. IEEE Trans. Computer C-22(7). Pp701-702, 1973. [11] E. Gardner and B. Derrida. Three unfinished works on the optimal storage capacity of networks. 3. Phys. A: Math. Gen., 22:1983-1994, 1989. [12] Thomas P. Trappenberg Fundamental of computational neuroscience, Oxford University Press, p190-196