Shigetaka Fujita. Rokkodai, Nada, Kobe 657, Japan. Haruhiko Nishimura. Yashiro-cho, Kato-gun, Hyogo , Japan. Abstract

Similar documents
Hopfield Neural Network and Associative Memory. Typical Myelinated Vertebrate Motoneuron (Wikipedia) Topic 3 Polymers and Neurons Lecture 5

7 Rate-Based Recurrent Networks of Threshold Neurons: Basis for Associative Memory

Neural Systems and Artificial Life Group, Institute of Psychology, National Research Council, Rome. Evolving Modular Architectures for Neural Networks

7 Recurrent Networks of Threshold (Binary) Neurons: Basis for Associative Memory

T sg. α c (0)= T=1/β. α c (T ) α=p/n

Stochastic Learning in a Neural Network with Adapting. Synapses. Istituto Nazionale di Fisica Nucleare, Sezione di Bari

Properties of Associative Memory Model with the β-th-order Synaptic Decay

0 o 1 i B C D 0/1 0/ /1

Data Mining Part 5. Prediction

Hierarchical Evolution of Neural Networks. The University of Texas at Austin. Austin, TX Technical Report AI January 1996.

of the dynamics. There is a competition between the capacity of the network and the stability of the

Logic Learning in Hopfield Networks

(a) (b) (c) Time Time. Time

Evolution of Adaptive Synapses: Robots with Fast Adaptive Behavior in New Environments

Artificial Intelligence Hopfield Networks

Structure Design of Neural Networks Using Genetic Algorithms

Improving the Performance of Evolutionary Optimization by. Dynamically Scaling the Evaluation Function.

Artificial Neural Network

1 Introduction Tasks like voice or face recognition are quite dicult to realize with conventional computer systems, even for the most powerful of them

Causality and communities in neural networks

Lecture 7 Artificial neural networks: Supervised learning

Using Variable Threshold to Increase Capacity in a Feedback Neural Network

Supervised Learning Part I

A. The Hopfield Network. III. Recurrent Neural Networks. Typical Artificial Neuron. Typical Artificial Neuron. Hopfield Network.

Learning and Memory in Neural Networks

Comparison of Evolutionary Methods for. Smoother Evolution. 2-4 Hikaridai, Seika-cho. Soraku-gun, Kyoto , JAPAN

Dynamical Systems and Deep Learning: Overview. Abbas Edalat

Evolutionary Computation

Processing of Time Series by Neural Circuits with Biologically Realistic Synaptic Dynamics

Individual learning and population evolution

High-conductance states in a mean-eld cortical network model

A gradient descent rule for spiking neurons emitting multiple spikes

A. The Hopfield Network. III. Recurrent Neural Networks. Typical Artificial Neuron. Typical Artificial Neuron. Hopfield Network.

Equivalence of Backpropagation and Contrastive Hebbian Learning in a Layered Network

fraction dt (0 < dt 1) from its present value to the goal net value: Net y (s) = Net y (s-1) + dt (GoalNet y (s) - Net y (s-1)) (2)

1 Introduction The Separation of Independent Sources (SIS) assumes that some unknown but independent temporal signals propagate through a mixing and/o

Neural Networks. Hopfield Nets and Auto Associators Fall 2017

Supporting Online Material for

Artificial Neural Network and Fuzzy Logic

Repeated Occurrences of the Baldwin Effect Can Guide Evolution on Rugged Fitness Landscapes

COMP304 Introduction to Neural Networks based on slides by:

Hopfield Neural Network

Public Key Exchange by Neural Networks

Evolutionary Games and Computer Simulations

Effects of Interactive Function Forms in a Self-Organized Critical Model Based on Neural Networks

2- AUTOASSOCIATIVE NET - The feedforward autoassociative net considered in this section is a special case of the heteroassociative net.

NEURAL NETWORK CHAOS ANALYSIS. A. Garliauskas. Mathematics and Informatics Institute, Akademijos 4, 2600 Vilnius, Lithuania.

Using a Hopfield Network: A Nuts and Bolts Approach

Probabilistic Models in Theoretical Neuroscience

Co-evolution of Morphology and Control for Roombots

Artificial Neural Networks Examination, March 2004

Biologically Inspired Compu4ng: Neural Computa4on. Lecture 5. Patricia A. Vargas

Evolution of Bilateral Symmetry in Agents Controlled by Spiking Neural Networks

In biological terms, memory refers to the ability of neural systems to store activity patterns and later recall them when required.

Neural Network Analysis of Russian Parliament Voting Patterns

CMSC 421: Neural Computation. Applications of Neural Networks

Integer weight training by differential evolution algorithms

Week 4: Hopfield Network

Inferring a System of Differential Equations for a Gene Regulatory Network by using Genetic Programming

Consider the way we are able to retrieve a pattern from a partial key as in Figure 10 1.

Hopfield Networks. (Excerpt from a Basic Course at IK 2008) Herbert Jaeger. Jacobs University Bremen

adap-org/ Jan 1994

Lecture 4: Feed Forward Neural Networks

Microcanonical Mean Field Annealing: A New Algorithm for Increasing the Convergence Speed of Mean Field Annealing.

Jer^ome Kodjabachian and Jean-Arcady Meyer. Ecole Normale Superieure. Abstract

Introduction. Previous work has shown that AER can also be used to construct largescale networks with arbitrary, configurable synaptic connectivity.

A Dynamical Implementation of Self-organizing Maps. Wolfgang Banzhaf. Department of Computer Science, University of Dortmund, Germany.

Jer^ome Kodjabachian and Jean-Arcady Meyer. Ecole Normale Superieure. Abstract

λ-universe: Introduction and Preliminary Study

Simple Model of Evolution. Reed College FAX (505) Abstract. of evolution. The model consists of a two-dimensional world with a

Using Assembler Encoding to Solve Predator- Prey Problem

in a Chaotic Neural Network distributed randomness of the input in each neuron or the weight in the

Storage capacity of hierarchically coupled associative memories

Artificial Neural Networks Examination, June 2005

THE UNIVERSITY OF SUSSEX. The Articial Evolution of Adaptive Behaviour. Inman Harvey. Submitted for the degree of D. Phil.

A Developmental Model for the Evolution of Complete Autonomous Agents

Keywords- Source coding, Huffman encoding, Artificial neural network, Multilayer perceptron, Backpropagation algorithm

EVOLUTIONARY OPERATORS FOR CONTINUOUS CONVEX PARAMETER SPACES. Zbigniew Michalewicz. Department of Computer Science, University of North Carolina

The Mixed States of Associative Memories Realize Unimodal Distribution of Dominance Durations in Multistable Perception

Asynchronous updating of threshold-coupled chaotic neurons

Computation with phase oscillators: an oscillatory perceptron model

Memories Associated with Single Neurons and Proximity Matrices

Effects of Interactive Function Forms and Refractoryperiod in a Self-Organized Critical Model Based on Neural Networks

arxiv: v2 [nlin.ao] 19 May 2015

DUAL-MODE DYNAMICS NEURAL NETWORKS FOR COMBINATORIAL OPTIMIZATION. Jun Park. A Dissertation Presented to the FACULTY OF THE GRADUATE SCHOOL

Evolutionary computation

Effects of refractory periods in the dynamics of a diluted neural network

A Novel Chaotic Neural Network Architecture

Artificial Neural Networks

Sorting Network Development Using Cellular Automata

Neural Networks. Textbook. Other Textbooks and Books. Course Info. (click on

Coevolution of learning and niche construction. Reiji Suzuki, Yasunori Noba and Takaya Arita

Artificial Neural Networks Examination, June 2004

Neuroevolution: from architectures to learning

CS:4420 Artificial Intelligence

Forming Neural Networks through Ecient and Adaptive. Coevolution. Abstract

arxiv:quant-ph/ v1 17 Oct 1995

Neural networks. Chapter 20. Chapter 20 1

Linear discriminant functions

Convolutional Associative Memory: FIR Filter Model of Synapse

Transcription:

KOBE-TH-94-07 HUIS-94-03 November 1994 An Evolutionary Approach to Associative Memory in Recurrent Neural Networks Shigetaka Fujita Graduate School of Science and Technology Kobe University Rokkodai, Nada, Kobe 657, Japan e-mail: fujitan@phys02.phys.kobe-u.ac.jp Haruhiko Nishimura Department of Information Science Hyogo University of Education Yashiro-cho, Kato-gun, Hyogo 673-14, Japan e-mail: HNISHIMU@JPNYITP.BITNET Abstract In this paper, we investigate the associative memory in recurrent neural networks, based on the model of evolving neural networks proposed by Nol, Miglino and Parisi. Experimentally developed network has highly asymmetric synaptic weights and dilute connections, quite dierent from those of the Hop- eld model. Some results on the eect of learning eciency on the evolution are also presented. keywords: Associative memory, Recurrent networks, Genetic algorithm, Evolution

1 Introduction As a model of associative memory in terms of the recurrent-type articial neural networks, Hopeld model [1] and its variants have been successfully developed and made some important results using the statistical mechanical techniques on the analogy with spin glass [2, 3]. Their assumption of the symmetrically and fully connected weights of synapses is, however, unrealistic, since this does not arise naturally in biological neural systems. Some researches are found allowing asymmetric synaptic weights or diluting the network connections [2, 3]. They do not concern the network structures, but most of them merely investigate the eects on the storage capacities under limited conditions. In any of above mentioned models, the physical positions where neurons exist or the distances between neurons are not considered at all. It must play an important role for the formation of biological networks in the brain. There should be the attempt to construct more natural models based on biologically and genetically plausible considerations, in which asymmetric and dilute connections are just the result. The genetic algorithm [4] can be a key ingredient toward such directions. Recently evolutionary search procedures have been combined with the articial neural networks (so called \the evolutionary articial neural networks") [5]. Almost all of the models deal with the feedforward multi-layered networks to raise the performance of the networks and in many of them the evolution of both weights and architectures is not integrated. In this paper we apply a genetic evolutionary scheme to recurrent neural networks and investigate associative memory. Aiming to construct a biologically inspired neural network, we take into account the following two points in network formation: The correspondence of genotype and phenotype should not be direct. Only developmental rules of neurons are encoded in the genotype representation. The information on the physical positions of neurons and the distances between neurons should be included in the model. The resulting network is a physical object in space. While the rst point, the indirect encoding scheme, has been studied [6, 7], few attempts have been made to incorporate both of the above two points, except for the model of \evolving (growing) neural network" by Nol, et al. [8, 9]. So we follow their scheme treating a neural network as a creature in a physical environment. The outline of our model is as follows. (1) Prepare the individuals which have the genotypes encoded the information for generating networks. (2) Learn the patterns according to an appropriate learning rule. (3) Impose the task for retrieving the given patterns. (4) Some individuals which make good results in the retrieval task are selected and inherit the genotypes to their ospring with mutation. (5) Repeat (2)-(4) until appropriate generation. As the generations go on, the abilities of the individuals will improve. 1

2 Model descriptions A population consists of M individuals. Each individual has a genotype encoded the information of creating the network structure. The genotype is divided into N blocks corresponding to N neurons. Each block contains the instructions for one neuron, namely, the \physical position" of the neuron ( x and y coordinates in the \physical space", 2 dimensional plane in this paper), the directional angle for the \axonal" growth, the axonal bifurcation angle, the axonal bifurcation length d, the \presynaptic" weight b (characterize the ring intensity and neuron's type, excitatory or inhibitory) and the \postsynaptic" weight a (amplify or damp the incoming signal) (Fig. 1). Note that these are essentially local information peculiar to each neuron. There is no direct information as to the connections between neurons. All values are randomly generated at the 0-th generation. 1 2 i N x i y i α i θ i d i b i a i Figure 1: A Genotype consisted of N blocks and parameters specied in each block. d α (x, y) θ From each neuron in the physical space, an axon grows and bifurcates as shown in Fig. 2 and connects with other neurons. The connections of neurons are determined by whether the axon reaches the neighbourhood r of other neurons or not. For convenience, we introduce the connectivity matrix cij ( i; j = 1 ; : : : ; N; cii = 0) as cij = 1 if the axon from the i-th neuron reaches the j-th neuron, and otherwise c = 0. Then the bare synaptic weight w0 at birth, before learning, is given by ij Figure 2: The growing and bifurcating process of an axon. ij 2

6 w0 0 ij = aibj if cij = 1, and wij = 0 if cij = 0. So the synaptic weights are generally asymmetric ( w0 0 ij = wji). After the bare weights are established, they are dressed by learning for a certain period of time. That is, p patterns fi = 61 g( = 1 ; : : : ; p) to be stored are N ew Old embedded into the network by an appropriate learning rule: wij = wij + wij. Iterated several times in one generation (learning task), the dressed weights are obtained. In general, the synaptic increment wij can be modulated in each learning step. Then the \retrieval task" is imposed to the neural network (phenotype) with the dressed weights wij. Concretely, one of p patterns with random noise is given to the individual network as the initial states of neurons. And the state of the neuron PN is updated synchronously according to the equation si( t + 1) = f( j=1 wijsj( t)), where si( t)( 01 si 1) is the state of the i-th neuron and t is the time step at retrieving. The transfer function f( x) is usually taken to be f( x) = tanh( x= 2 ). During the update proceeds, we trace the overlap of the state of the network with P N the target pattern: m ( t ) = 1 =N i=1 i si( t). This procedure is applied to all p patterns. The retrieval task is repeated several times with changing the noise of the set of p patterns, because the proper network should have the balanced robustness against the noise. For this reason, we take the tness as the sum of the values of overlaps over all tasks. At the end of one generation, some individuals which have got higher tness values are selected. Their genotypes are inherited to their ospring with random mutation (we do not use crossover for simplicity). The above process continues for appropriate generations. 3 Simulations and results To carry out the experiments, we have made the following choices of parameters. A population consists of 50 ( M = 50) individuals. The best 10 of individuals which have got higher tness values are selected, and 5 copies of each of their genotypes are inherited to their ospring with mutation. The mutation rate is 0.004. We take a square, normalized to 121, as the physical space. The ranges of each parameter for a genotype are 0 x; y 1, 0 2, 0 = 3, 0 d 0: 15, 02 b 2 and 0 a 2. The axon bifurcates 5 times and the neighbourhood of a neuron is xed to r = 0: 05. And the steepness parameter in the transfer function is set to 0.015. In learning and retrieval tasks, we use the 727 non-orthogonal 4 ( p = 4) patterns as a set of target patterns (Fig. 3), then the total number of neurons becomes 49 ( N = 49). As the learning rule we here introduce simple Hebbian rule for the connected Pp neurons: wij = =1 i j if cij = 1, and wij = 0 if cij = 0, where is the "learning eciency" represents the intensity of a posteriori learning after the birth. Note that we do not take into consideration the evolution of learning rule itself. In this paper we make xed through generations. Learning and retrieval task processes in one generation are described by Fig. 4. First,15 steps are given for 3

Figure 3: An example of a set of target patterns. RT1 RT2 RT3 RT4 Learning Task Retrival Tasks 1 Generation Figure 4: Flow of learning and retrieval task processes in one generation. learning. Secondly, for each of 4 patterns 15 steps are assigned to retrieving trial by starting with the noisy pattern state with 5% random noise. Then the overlaps are marked at the step t=5,10 and 15 and summed up. The retrieval task is repeated 4 times successively (RT1RT4) by changing the noise. Bold solid lines in Fig. 5 show the transition of the tness of the best individual until the 1000th generation at = 0: 002 (Fig. 5(a)) and = 0: 05 (Fig. 5(b)), respectively. The tness value is normalized to max 100. The \bare weight" line(thin) represent the ability of the bare weights of the best individual, made by removing the learned changes from the dressed weights, namely the \native ability" for retrieving. The \pure genetic" line(dashed) represents the ability of genetic algorithm only without learning after the birth ( = 0), namely means acquisition of the \inherited memory". Changing the learning eciency, we found that higher the parameter becomes, more the native ability goes down. It appears that the presence of learning helps the network to t its environment and makes evolution easier. Fig. 6 and Fig. 7 show the network structures of the best individual and the distributions of distance versus weight value between connected neurons generated at the 0th generation(g0) and the 1000th(G1000) in the case of = 0: 05, respectively. Here we cite the comparison of our numerical results with those of the conventional Hopeld model. In the following, H indicates the Hopeld model and E indicates the best individual of our evolutionary model of = 0: 05 at the 1000th generation. The total numbers of connections are H:2352 and E:406, then the mean values of the number of connections per neuron become H:48 and E:8.29. This shows E develops strongly diluted connections. The values of symmetry parameter, P P dened by = w w = w2, are given by H:1.0 and E:0.23. This means E i;j ij ji i;j ij 4

1000 (a) 1000 (b) Dressed Weight 800 Pure Genetic 800 Fitness 600 400 Dressed Weight Fitness 600 400 Pure Genetic Bare Weight 200 Bare Weight 200 0 0 200 400 600 800 1000 G(eneration) 0 0 200 400 600 800 1000 G Figure 5: Evolution of tness of the best individual in population. (a) = 0: 002. (b) = 0: 05. G0 G1000 Figure 6: Network structures of the best individual at G0 and G1000 ( =0.05). forms highly asymmetric weights. Finally we mention about the tness values. E can reach 100 tness value for noisy(5%) patterns, but H cannot reach 100 and gets 88.3 in average naively (remember that we admit non-orthogonal patterns). And E is almost stable for suf- cient retrieving time. This suggests that our evolutionary approach may exceed the Hopeld model in associative memory, in spite of quite dierent structure of networks. 4 Summary We have presented the results of simulations of associative memory in the recurrenttype neural networks composed of growing neurons. We found that the developed networks are far dierent from the Hopeld model in respect of the symmetricity 5

Figure 7: Dressed weight value versus distance between connected neurons of the best individual at G0 and G1000 ( =0.05). and the connection number per neuron of the connection weights, but it is possible to act as a sucient or a better associative memory than the conventional Hopeld model. In this paper we kept our model simple, in terms of minimizing computational requirements and studying the feasibility of the methodology of biologically inspired associative memory models. Further simulations are going to concentrate on the inuence of increasing the sets of task-patterns. It is expected that the dynamic environment where diverse tasks are present excludes the task-specicity and makes us possible to examine the relationship between learning and evolution [10, 11]. It seems interesting as a future direction of research to incorporate epigenetic factors [12] such as the primary processes of neuronal development (adhesion, apoptosis etc.) and/or the action of nerve growth factor from target neurons. 6

References [1] J.J. Hopeld. Neural Networks and Physical Systems with Emergent Collective Computational Abilities, Proc. Natl. Acad. Sci. USA 79, pp.2554-2558, 1982. [2] B. Muller and J. Reinhardt. Neural Networks: An Introduction, Springer- Verlag, 1990. Introduction to the Theory of Neural Com- [3] J. Hertz, A. Krogh and R. Palmer. putation, Addison-Wesley, 1991. [4] J.H. Holland. Adaptation in Natural and Articial Systems, University of Michigan Press, 1975. [5] X. Yao. Evolutionary Articial Neural Networks, Int. J. Neural Systems Vol. 4, pp.203-222, 1993. [6] H. Kitano. Designing Neural Networks using Genetic Algorithms with Graph Generation System, Complex Systems 4, pp.461-476, 1990. [7] F. Gruau. Genetic synthesis of boolean neural networks with a cell rewriting developmental processes, in Proc. of the Int. Workshop on Combinations of Genetic Algor. and Neural Networks (COGANN-92), eds. D. Whitley and J. D. Schaer, IEEE Computer Society Press, Los Alamitos, CA, pp. 55-74, 1992. [8] S. Nol and D. Parisi. Growing Neural Networks, Technical Report 91-15, Institute of Psychology C.N.R. - Rome, 1991. PCIA- [9] S. Nol, O. Miglino and D. Parisi. Phenotypic Plasticity in Evolving Neural Networks, Technical Report PCIA-94-05, Institute of Psychology C.N.R. - Rome, 1994. [10] G. Hinton and S. Nowlan. How Learning Can Guide Evolution, Complex Systems 1, pp.495-502, 1987. [11] D. Ackley and H. Littman. Interactions Between Learning and Evolution, in Articial Life II, eds. C.G. Langton, C. Taylor, J.D. Farmer and S. Rasmussen, Addison-Wesley, pp. 487-509, 1991. [12] G. Edelman. Neural Darwinism: The Theory of Neuronal Group Selection, Basic Books, 1987. 7