Terminal attractor optical associative memory with adaptive control parameter

Similar documents
Storage Capacity of Letter Recognition in Hopfield Networks

Hopfield Neural Network and Associative Memory. Typical Myelinated Vertebrate Motoneuron (Wikipedia) Topic 3 Polymers and Neurons Lecture 5

CHAPTER 3. Pattern Association. Neural Networks

Artificial Intelligence Hopfield Networks

Hopfield Network for Associative Memory

3.3 Discrete Hopfield Net An iterative autoassociative net similar to the nets described in the previous sections has been developed by Hopfield

Using Variable Threshold to Increase Capacity in a Feedback Neural Network

A. The Hopfield Network. III. Recurrent Neural Networks. Typical Artificial Neuron. Typical Artificial Neuron. Hopfield Network.

Consider the way we are able to retrieve a pattern from a partial key as in Figure 10 1.

( ) T. Reading. Lecture 22. Definition of Covariance. Imprinting Multiple Patterns. Characteristics of Hopfield Memory

Using a Hopfield Network: A Nuts and Bolts Approach

A. The Hopfield Network. III. Recurrent Neural Networks. Typical Artificial Neuron. Typical Artificial Neuron. Hopfield Network.

CSE 5526: Introduction to Neural Networks Hopfield Network for Associative Memory

HOPFIELD neural networks (HNNs) are a class of nonlinear

Nonlinearity optimization in nonlinear joint transform correlators

Associative Memory : Soft Computing Course Lecture 21 24, notes, slides RC Chakraborty, Aug.

Computational Intelligence Lecture 6: Associative Memory

In biological terms, memory refers to the ability of neural systems to store activity patterns and later recall them when required.

Synchronous vs asynchronous behavior of Hopfield's CAM neural net

SCRAM: Statistically Converging Recurrent Associative Memory

Iterative Autoassociative Net: Bidirectional Associative Memory

Bobby Hunt, Mariappan S. Nadar, Paul Keller, Eric VonColln, and Anupam Goyal III. ASSOCIATIVE RECALL BY A POLYNOMIAL MAPPING

Hertz, Krogh, Palmer: Introduction to the Theory of Neural Computation. Addison-Wesley Publishing Company (1991). (v ji (1 x i ) + (1 v ji )x i )

Neural Networks. Associative memory 12/30/2015. Associative memories. Associative memories

Neural Networks Lecture 6: Associative Memory II

Error analysis in Barker s method: extension to ternary systems

Lecture 7 Artificial neural networks: Supervised learning

Neural Nets and Symbolic Reasoning Hopfield Networks

Hopfield Neural Network

Application of hopfield network in improvement of fingerprint recognition process Mahmoud Alborzi 1, Abbas Toloie- Eshlaghy 1 and Dena Bazazian 2

A Remark on Alan H. Kawamoto: Nonlinear Dynamics in the Resolution of Lexical Ambiguity: A Parallel Distributed Processing Account

Week 4: Hopfield Network

7 Rate-Based Recurrent Networks of Threshold Neurons: Basis for Associative Memory

7 Recurrent Networks of Threshold (Binary) Neurons: Basis for Associative Memory

2- AUTOASSOCIATIVE NET - The feedforward autoassociative net considered in this section is a special case of the heteroassociative net.

Optical implementation of the Hopfield model

Memories Associated with Single Neurons and Proximity Matrices

Learning and Memory in Neural Networks

The Mixed States of Associative Memories Realize Unimodal Distribution of Dominance Durations in Multistable Perception

Properties of Associative Memory Model with the β-th-order Synaptic Decay

How can ideas from quantum computing improve or speed up neuromorphic models of computation?

On the SIR s ( Signal -to- Interference -Ratio) in. Discrete-Time Autonomous Linear Networks

Computational Intelligence Lecture 3: Simple Neural Networks for Pattern Classification

ARTIFICIAL NEURAL NETWORK PART I HANIEH BORHANAZAD

arxiv: v2 [nlin.ao] 19 May 2015

Large Fluctuations in Chaotic Systems

THE information capacity is one of the most important

Intelligent Modular Neural Network for Dynamic System Parameter Estimation

4. Multilayer Perceptrons

Pattern Association or Associative Networks. Jugal Kalita University of Colorado at Colorado Springs

Convolutional Associative Memory: FIR Filter Model of Synapse

Neural networks. Chapter 20. Chapter 20 1

Nonlinear Prediction for Top and Bottom Values of Time Series

IN THIS PAPER, we consider a class of continuous-time recurrent

Artificial Neural Networks Examination, March 2004

Neural networks. Chapter 19, Sections 1 5 1

Neural Networks Lecture 2:Single Layer Classifiers

Lecture 6. Notes on Linear Algebra. Perceptron

COMP9444 Neural Networks and Deep Learning 11. Boltzmann Machines. COMP9444 c Alan Blair, 2017

Supporting Online Material for

PULSE-COUPLED networks (PCNs) of integrate-and-fire

Error Correcting Capability of Cellular Automata Based Associative Memory

Neural Networks. Prof. Dr. Rudolf Kruse. Computational Intelligence Group Faculty for Computer Science

Neural Nets in PR. Pattern Recognition XII. Michal Haindl. Outline. Neural Nets in PR 2

Neural networks: Unsupervised learning

Data Mining Part 5. Prediction

Neural Networks Introduction

CSC321 Lecture 4 The Perceptron Algorithm

T sg. α c (0)= T=1/β. α c (T ) α=p/n

Handout 2: Invariant Sets and Stability

Laplace Transform Analysis of Signals and Systems

Chapter 9: The Perceptron

Pattern Recognition System with Top-Down Process of Mental Rotation

An Introductory Course in Computational Neuroscience

HOPFIELD MODEL OF NEURAL NETWORK

Artificial Neural Networks

Artificial Neural Networks Examination, June 2005

Gibbs function continuation for the stable computation of chemical equilibrium

Neural Network Analysis of Russian Parliament Voting Patterns

Chapter 1 Error Analysis

K. Pyragas* Semiconductor Physics Institute, LT-2600 Vilnius, Lithuania Received 19 March 1998

CS 4700: Foundations of Artificial Intelligence

COMS 4771 Introduction to Machine Learning. Nakul Verma

Introduction to Artificial Neural Networks

Neural networks. Chapter 20, Section 5 1

Lecture 4: Perceptrons and Multilayer Perceptrons

Determining the Optimal Decision Delay Parameter for a Linear Equalizer

Frequency Selective Surface Design Based on Iterative Inversion of Neural Networks

Branislav K. Nikolić

Associative Memories (I) Hopfield Networks

Stochastic Model for Adaptation Using Basin Hopping Dynamics

Keywords- Source coding, Huffman encoding, Artificial neural network, Multilayer perceptron, Backpropagation algorithm

Introduction To Artificial Neural Networks

Neural Networks for Machine Learning. Lecture 11a Hopfield Nets

In-Memory Computing of Akers Logic Array

Storage capacity of hierarchically coupled associative memories

Revision: Neural Network

Noise-to-State Stability for Stochastic Hopfield Neural Networks with Delays

Introduction to Neural Networks

Delsarte s linear programming bound

Transcription:

1 June 1998 Ž. Optics Communications 151 1998 353 365 Full length article Terminal attractor optical associative memory with adaptive control parameter Xin Lin a,), Junji Ohtsubo b, Masahiko Mori a a Electrotechnical Laboratory, 1-1-4 Umezono, Tsukuba 305, Japan b Faculty of Engineering, Shizuoka UniÕersity, 3-5-1 Johoku, Hamamatsu 432, Japan Received 4 November 1997; accepted 27 January 1998 Abstract Ž. The control parameter dependence in terminal attractor TA optical associative memory is investigated. With adaptive setting of the control parameter in the dynamics of the TA model for dynamic iteration, perfect convergence of the associative memory for pattern recognition is achieved. With numerical simulations, the optimal control parameter in the TA model associative memory is determined. The optimal control parameter is also used in an optical experiment. The experimental results show that the target pattern can be successfully associated. q 1998 Elsevier Science B.V. All rights reserved. Keywords: Optical associative memory; Terminal attractor; Control parameter 1. Introduction One of the major research topics of neural networks is in the area of associative memory. Hopfield described a simple neural network model for the operation of associawx 1. The action of tive model and pattern recognition individual neurons is modeled through a thresholding operation and the information is stored in stable equilibrium points in the interconnections among the neurons. The computation is performed by setting the state Ž on or off. of some of the neurons according to external stimulus and, with the interconnected according to the recipe that Hopfield prescribed, the state of all neurons that are interconnected to those that are externally stimulated spontaneously converge to the stored pattern that is the most similar to the external input. The basic operation per- ) E-mail: xlin@etl.go.jp formed is a nearest-neighbor search, a fundamental operation for pattern recognition, associative memory, and error correction. But further investigation reveals that the basins of the attractors for stored patterns are small in the Hopfield neural network and the recalling ability is not so great wx 2. The existence of oscillating states also affects memory capacity wx 3. For the elimination of the spurious state in the Hopfield associative memory, a type of attractor called a terminal attractor Ž TA., which represents singular solutions of a neural dynamic system, was introduced wx 4. These terminal attractors are characterized by having finite relaxation times, no spurious states, and infinite stability. Applications of terminal attractors for content-addressable and associative memories, pattern recognition, self-organization, and for dynamical training have been illustrated. Recently, for the purpose of comparing the Hopfield model both including and excluding the terminal attractor, an optical neural network with terminal attractors for pattern recognition is proposed wx 5. The results show that the spurious states of the energy function in the Hopfield 0030-4018r98r$19.00 q 1998 Elsevier Science B.V. All rights reserved. Ž. PII S0030-4018 98 00059-5

354 ( ) X. Lin et al.roptics Communications 151 1998 353 365 Fig. 1. Stored patterns in the network. associative memory can be avoided and the recognition rate is considerably enhanced. The capacity of TA model associative memory has been also investigated based on the consistency between the stored pattern and the obtained equilibrium state in statistical thermodynamics by numerical simulations wx 6. The absolute capacity of the TA model is greater than 0.35N, which contrasts with the relative capacity or the theoretical absolute capacity for the conventional associative memory. However, in the TA model associative memory dynamic system, the control parameters are significant. In this paper, we investigate the control parameter dependence in TA optical associative memory. First, we determine the optimal control parameter for pattern recognition by the numerical simulations. Then, the optimal control parameter is used in an optical experiment. The experimental results show that the target pattern can be successful associated. 2. Theory 2.1. TA model for associatiõe memory model for associative memory w4,5x should be briefly reviewed. Let us consider a discrete time type of neural network with N neurons and M stored patterns. The ith component of the state vector xi at time tq1 may be written as N Ý x Ž tq1. s W f x Ž t. i ij j js1 M 1r3 y Ý a fwxiž. t xyxi 4 ms 1 2 ½ w i x i 4 5 = exp yb f x Ž. t yx, Ž. 1 where the x 4Ž ms1,2,..., M. i is a set of M linearly independent stored vectors, x Ž. i t is the output of the ith neuron at time t, Wij is a synaptic matrix determined by Hebb s law in the same way as the Hopfield model, and a and b are positive control constants. In this model, we select the value of the parameter a as unity for all m from empirical basis. Then, Eq. Ž. 1 is written as Table 2 Hamming distances between inputs and stored patterns Before the control parameter dependence in TA optical associative memory is discussed, the dynamics of the TA Table 1 Hamming distances among stored patterns A B E H K M N O R S X Y A 0 40 35 27 37 25 27 27 40 37 35 41 B 0 15 33 39 37 37 27 16 19 35 37 E 0 36 32 42 40 30 19 26 28 32 H 0 24 14 20 24 35 32 32 38 K 0 28 30 40 29 36 22 26 M 0 16 30 37 30 30 32 N 0 34 39 32 22 30 O 0 33 24 42 42 R 0 29 23 35 S 0 38 36 X 0 16 Y 0 Mean Hamming distance Hs30.8

( ) X. Lin et al.roptics Communications 151 1998 353 365 355 Fig. 2. Hamming distance H between output patterns and stored pattern Y versus the iteration time t in the original Ž. a, and the modified Ž. TA associative memory b. Ns100, b s0.2.

356 ( ) X. Lin et al.roptics Communications 151 1998 353 365 Fig. 3. Hamming distance H between output patterns and stored pattern Y versus the iteration time t in the original Ž. a, and the modified Ž. TA associative memory b. Ns100, b s0.5.

( ) X. Lin et al.roptics Communications 151 1998 353 365 357 Fig. 4. Hamming distance H between output patterns and stored pattern Y versus the iteration time t in the original Ž. a, and the modified Ž. TA associative memory b. Ns100, b s1.

358 ( ) X. Lin et al.roptics Communications 151 1998 353 365 Fig. 5. Hamming distance H between output patterns and stored pattern Y versus the iteration time t in the original Ž. a, and the modified Ž. TA associative memory b. Ns100, b s3.

( ) X. Lin et al.roptics Communications 151 1998 353 365 359 N Ý x Ž tq1. s W f x Ž t. i ij j js1 M 1r3 Ý w iž. x i 4 ms 1 2 ½ w i x i 4 5 y f x t yx =exp yb f x Ž. t yx, Ž. 2 where the function f wxis a neuron nonlinear function and written as fwxiž. t xstanhwxiž. t x. Ž. 3 That is, the network changes its state at discrete ts 0,1,2,... according to Eqs. Ž. 2 and Ž. 3. Here, the nonlinear threshold function tanhwxoperates componentwise on vectors. 2.2. TA model for optical associatiõe memory The optical implementation of the dynamics that is described by Eq. Ž. 2 is not easy because it contains a square and a 1r3 power function. For the feasibility of the optical implementation of the TA model associative memory, we make further approximations to Eqs. Ž. 2 and Ž. 3 without losing the essence of the TA dynamics system. If we assume unipolar binary number Ž 1, 0. for the neuronstate vectors, the factor 1r3 in the power function of Eq. Ž. 2 may be dropped and, instead of the square operation in the exponential function, the absolute value can be used without changing the value of the dynamics system. Based on the property of unipolar representation of this binary system, Eqs. Ž. 2 and Ž. 3 can be rewritten as wx 5 N Ý x Ž. t s W f x Ž. t y f x Ž. t yx M Ý w x 4 i ij j i i js1 ms1 =expyb fwxiž. t xyx i 4, Ž. 4 Ž. w Ž. x Ž. f x t s1 x t, 5 j i where 1wuxs1 when u)0 and y1 when u-0. Eqs. Ž. 2 and Ž. 3 will be called original, Eqs. Ž. 4 and Ž. 5 will be called modified TA model, respectively hereafter. 3. Computer simulation 3.1. Method In general, for associative memory for pattern recognition, the convergence of the iteration not only depends on the data representation, the number and the content of the stored vectors, the initial input, the synapses, and the thresholding value, but also depends on the control param- eter b in the exponential term of Eqs. Ž. 2 and Ž. 4. The Fig. 6. Examples of the recalling process in the original TA model. The Hamming distance of the input from the stored pattern Y is 5. t is the iteration time.

360 ( ) X. Lin et al.roptics Communications 151 1998 353 365 purpose of b in the terminal attractor is to provide a w Ž.x Gaussian distribution peaked at f xi t sx i. In prac- tice, the value of b controls the influence of the exponential term at the neighborhood of the stored pat- Ž terns. A small b for instance, b - 1. causes stronger cross-talk among stored patterns, and the spurious states caused by cross talks are generated near the boundary of a basin of a stored pattern. Whereas, a large b reduces the cross talk. In other words, the value of b decides the behavior of neighbors of a terminal attractor. Unfortunately, it is quite difficult to determine the b value based on a rigorous analytical method. Therefore, with the assistance of numerical simulation, the trial-and-error method will be used for selection of the control parameter b. Numerical simulations have been performed by using a 10=10 neurons network based on Eqs. Ž. 2 and Ž. 3. Twelve characters as shown in Fig. 1 are embedded as terminal attractors in the network. The Hamming distance between the pattern X 1 and X 2 is defined by p q 1 2 < 1 2 ÝÝ ij ij < isj js1 HŽ X, X. s X yx Ž pqsn.. Ž 6. By Eq. Ž. 6, the Hamming distances from each stored pattern are tabulated in Table 1. The smallest Hamming distance is 14, and the mean Hamming distance among 12 patterns is 30.8. Four initial imperfect input patterns and the Hamming distances of these inputs from stored patterns are tabulated in Table 2. The Hamming distance of the imperfect inputs from stored pattern Y are 3, 5, 7, and 9, respectively, and those for the other stored patterns are equal to 17 or more. 3.2. Results and discussions We tested four parameter b values of 0.2, 0.5, 1 and 3 using the above 10=10 neurons network. The results for the recalling property of the TA model associative memory are shown in Figs. 2, 3, 4, and 5. In these figures, the abscissas are the iteration time and the ordinates are the Hamming distance of the recalled pattern from a stored pattern Y. Figures Ž. a and Ž. b correspond to the original and modified TA model dynamics system, respectively. Fig. 2 shows the case of b s0.2. When the Hamming distances between the input patterns and stored pattern are 7 and 9, even if the recollections are unsuccessful, spurious states are not recalled, instead, both networks show unstable chaotic states, i.e., without reaching any Fig. 7. Examples of the recalling process in the original TA model. The Hamming distance of the input from the stored pattern Y is 9. t is the iteration time.

( ) X. Lin et al.roptics Communications 151 1998 353 365 361 equilibrium. When the Hamming distance is 5, the recalled patterns are trapped to spurious states even if the networks are stable. Only for the case of Hamming distance equal to 3, the correct patterns are recalled without error. This indicates that the basins of attraction of stored patterns are small and the recollection ability is not very large if b is 0.2. In Fig. 3, b is assumed to be 0.5. We can see that when the Hamming distances are 7 and 9, the networks are trapped to equilibrium states that are very different from the stored pattern. These results in Figs. 2 and 3 indicate that a small b will generate unstable oscillation or spurious states near the boundaries of the basin of the stored patterns when the Hamming distance between the input pattern and stored target pattern is greater due to cross talks among the stored patterns. Figs. 4 and 5 correspond to b s1 and 3, respectively. Both the two networks converge to the correct pattern within a finite iteration time. However, we can see that for the network with b s3 the recalling speed is lower than for the network with b s1. Some instant states in the evolution of the networks with b s0.2, 0.5, 1, 3 are show in Figs. 6 9. t is the iteration time. Figs. 6 and 7 are for the original TA model dynamics system and, the Hamming distances of the inputs from the target pattern Y are 5 and 9, respectively. When b s1 and 3, both Figs. 6 and 7, the networks correctly converged to Y starting from imperfect patterns. When b s0.2, the case of Fig. 6 traps to a stable spurious pattern and, Fig. 7 shows unstable oscillation states until iteration time of 100 or larger, to be brief, when b is smaller, the recollection of the associative memory is unsuccessful in most cases. Figs. 8 and 9 correspond to the modified TA model dynamics system. We can see that the obtained results are the same as for the original TA model in Figs. 6 and 7. From these numerical simulation results, Ž a large b b G1. can reduce the cross talk and the dynamics system exhibits higher recollection ability and stable recalling property. However, these results are only for four test values and for a specific neural network Ž 10=10 neurons.. In order to determine the optimal values of the control parameter b for various neural networks, more b values are tested. Fig. 10 shows the recalling ability for various b. The network size is 10=10. We can see that the recalling ability increases with increasing b. But, after b reaches 1, the recalling ability remains at same value. We also use four b values to test the recalling ability for Fig. 8. Examples of the recalling process in the modified TA model. The Hamming distance of the input from the stored pattern Y is 5. t is the iteration time.

362 ( ) X. Lin et al.roptics Communications 151 1998 353 365 Fig. 9. Examples of the recalling process in the modified TA model. The Hamming distance of the input from the stored pattern Y is 9. t is the iteration time. Fig. 10. The recalling ability versus b values.

( ) X. Lin et al.roptics Communications 151 1998 353 365 363 Fig. 11. The recalling ability versus network size. various network sizes, as shown in Fig. 11. We can see that the network size almost does not influence the choice of the b value. 4. Optical implementation 4.1. Method Optical implementations have been performed for pattern recognition with the proposed adaptive control parameter. The optical experimental system uses liquid-crystal television Ž LCTV. spatial light modulators Ž SLMs. and CCD cameras based on the modified TA model dynamics system of Eqs. Ž. 4 and Ž. 5 as shown in Fig. 12. It consists of two subsystems. One is the optical system for the Hopfield model and the other is the optical system which realizes the TA model. LCTV 1 and LCTV 2 are used to represent the state vector fw x Ž.x i t and the connectivity matrix W ij, respectively. For implementation of the optical subtraction, we used a real-time image subtraction system by using polarization modulation of LCTV ŽLCTV3 and LCTV4. wx 7. The exponential operation realized by the optical subsystem that has light transmission property equal to the value of the exponential function, which is a fader control is used. The signals from the Hopfield subsystem and the TA subsystem are detected by the CCD camera 1 and 4, respectively, and are sent to the computer to calculate the difference between them and the next state vector fw x Ž tq1.x. The output fwx Ž tq1.x i i is used as the input of a new state to the LCTV-SLM for the dynamical system. This process is iterated until convergence is reached. The converged state is displayed on a TV monitor. A detailed discussion on this optical implementation wx can be found in Ref. 5. 4.2. Results For the feasibility of the optical implementation, we used a network with 4=4 neurons and choose the value of b as 1 for all m. The network was trained to recognize the patterns A, O, and H. Fig. 13 shows an example of the experimental results of the recovery from imperfect inputs by the optical setup in Fig. 12. Three stored patterns are shown in Fig. 13Ž. a. The mean Hamming distance among these stored patterns is 8. Fig. 13Ž. b shows four initial imperfect patterns. The Hamming distances of the inputs from the target pattern A are 1, 2, 3, and 4 from top to bottom, respectively, and those with the other stored patterns are equal to 5 or larger. The Hamming distances of the input patterns from the stored patterns are tabulated in Table 3. The output results are shown in Fig. 13Ž. c. We can see that the network correctly converges to A, thus the association is successful for all input data. These experimental results indicate that the optical results are Table 3 Hamming distances of imperfect patterns from the stored patterns Input pattern Stored pattern Ž1. Ž2. Ž3. x A x O x H X 1 5 7 1 X 2 6 10 2 X3 3 7 11 X4 4 6 12

364 ( ) X. Lin et al.roptics Communications 151 1998 353 365 Fig. 12. Optical TA neural network. P1 P6: polarizers. Table 4 Parameter b dependence in the TA model associative memory b -1 b s1 b )1 Crosstalk between stored strong reduced reduced patterns Final state of the network trapping to spurious states converging to right converging to right unstable chaotic state stable state stable state Recalling ability of the network bad good good Recalling speed of the network fast low Optical implementation of the optical adjustment is easy; optical adjustment is difficult; network output image quality is good noise is strong

( ) X. Lin et al.roptics Communications 151 1998 353 365 365 Ž. Ž. Ž. Fig. 13. Experiment results. a Stored patterns, b inputs, c outputs with b s1. consistent with the numerical simulation even if a different network Ž with 4=4 neurons. from the simulation Žwith 10=10 neurons. is used. good. In practice, we have been performed an optical implementation using a network with 4=4 neurons, three stored patterns, and b s1. Correct retrieval is demonstrated by the experimental results. 5. Conclusion We have given the control parameter b dependence in the TA model associative memory. The numerical simulations for pattern recognition with different values of b were performed by using different network models. The results are summarized in Table 4. We can see that when b is less than 1, the network becomes unstable and recalling fails. Otherwise, when b is greater than or equal to 1, the network shows good stability and correctly converges to the desired pattern. Specifically, when b is equal to 1, the recalling speed is also fast, and optical adjustability and output image quality are References wx 1 J.J. Hopfield, Proc. Natl. Acad. Sci. U.S.A. 79 Ž 1982. 2254. wx 2 R.J. McEliece, E.C. Posner, E.R. Rodemich, S.S. Venkatesh, IEEE Trans. Inf. Theory IT 33 Ž 1987. 461. wx 3 B.L. Montgomery, B.V.K. Vijaya Kumar, Appl. Optics 25 Ž 1986. 3759. wx 4 M. Zak, Neural Net. 2 Ž 1989. 259. wx 5 X. Lin, J. Ohtsubo, Opt. Eng. 35 Ž.Ž 8 1996. 2145. wx 6 X. Lin, J. Ohtsubo, Optics Comm. 146 Ž 1998. 49. wx 7 H. Sakai, J. Ohtsubo, Appl. Optics 31 Ž 1992. 6852.