<f> Generation t. <f> Generation t

Similar documents
Statistical Dynamics of the Royal Road Genetic Algorithm

(a) (b) (c) (d) N = 8 K = 8 M = 200 q = 5 x N = 6 K = 6 M = 150 q = 1.6 x N = 4 K = 10 M = 150 q = 1.4 x 10-2

Exploration of population fixed-points versus mutation rates for functions of unitation

Complex Systems Theory and Evolution

The Evolutionary Unfolding of Complexity

adap-org/ Jan 1994

Quotient Evolutionary Space: Abstraction of Evolutionary process w.r.t macroscopic properties

Evolutionary Computation

Effects of Neutral Selection on the Evolution of Molecular Species

A Method for Estimating Mean First-passage Time in Genetic Algorithms

Dynamical Embodiments of Computation in Cognitive Processes James P. Crutcheld Physics Department, University of California, Berkeley, CA a

Introduction. Genetic Algorithm Theory. Overview of tutorial. The Simple Genetic Algorithm. Jonathan E. Rowe

arxiv:physics/ v1 [physics.bio-ph] 27 Jun 2001

Shigetaka Fujita. Rokkodai, Nada, Kobe 657, Japan. Haruhiko Nishimura. Yashiro-cho, Kato-gun, Hyogo , Japan. Abstract

Looking Under the EA Hood with Price s Equation

Statistical Complexity of Simple 1D Spin Systems

EVOLUTIONARY OPERATORS FOR CONTINUOUS CONVEX PARAMETER SPACES. Zbigniew Michalewicz. Department of Computer Science, University of North Carolina

The Role of Crossover in Genetic Algorithms to Solve Optimization of a Function Problem Falih Hassan

Evolutionary Computation Theory. Jun He School of Computer Science University of Birmingham Web: jxh

cells [20]. CAs exhibit three notable features, namely massive parallelism, locality of cellular interactions, and simplicity of basic components (cel


Mechanisms of Emergent Computation in Cellular Automata

A Statistical Genetic Algorithm

Implicit Formae in Genetic Algorithms

Revisiting the Edge of Chaos: Evolving Cellular Automata to Perform. Santa Fe Institute Working Paper (Submitted to Complex Systems)

Optimizing Stochastic and Multiple Fitness Functions

On the convergence rates of genetic algorithms

A New Approach to Estimating the Expected First Hitting Time of Evolutionary Algorithms

Learning with Ensembles: How. over-tting can be useful. Anders Krogh Copenhagen, Denmark. Abstract

A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring

Genetic Algorithms with Dynamic Niche Sharing. for Multimodal Function Optimization. at Urbana-Champaign. IlliGAL Report No

(a) (b) (c) Time Time. Time

Average Reward Parameters

Form Invariance and Implicit Parallelism

Rule table φ: Lattice: r = 1. t = 0. t = 1. neighborhood η: output bit: Neighborhood η

Haploid-Diploid Algorithms

A comparison of evolutionary and coevolutionary search

Bistability of the Needle Function in the Presence of Truncation Selection

Simple Model of Evolution. Reed College FAX (505) Abstract. of evolution. The model consists of a two-dimensional world with a

Genetic algorithms: bridging the convergence gap

AN EVOLUTIONARY ALGORITHM TO ESTIMATE UNKNOWN HEAT FLUX IN A ONE- DIMENSIONAL INVERSE HEAT CONDUCTION PROBLEM

0 o 1 i B C D 0/1 0/ /1

Evolutionary Computation. DEIS-Cesena Alma Mater Studiorum Università di Bologna Cesena (Italia)

Evolving on a Single-peak Fitness Landscape. Lionel Barnett. Centre for the Study of Evolution. University of Sussex.

Brazilian Journal of Physics, vol. 27, no. 4, december, with Aperiodic Interactions. Instituto de Fsica, Universidade de S~ao Paulo

550 XU Hai-Bo, WANG Guang-Rui, and CHEN Shi-Gang Vol. 37 the denition of the domain. The map is a generalization of the standard map for which (J) = J

Evolutionary Games and Computer Simulations

The Bias-Variance dilemma of the Monte Carlo. method. Technion - Israel Institute of Technology, Technion City, Haifa 32000, Israel

N.G.Bean, D.A.Green and P.G.Taylor. University of Adelaide. Adelaide. Abstract. process of an MMPP/M/1 queue is not a MAP unless the queue is a

Gene Pool Recombination in Genetic Algorithms

Proceedings of the Seventh Oklahoma Conference on Articial Intelligence, pp , November, Objective Function

What Makes a Problem Hard for a Genetic Algorithm? Some Anomalous Results and Their Explanation

Scalability of Selectorecombinative Genetic Algorithms for Problems with Tight Linkage

below, kernel PCA Eigenvectors, and linear combinations thereof. For the cases where the pre-image does exist, we can provide a means of constructing

Improving the Performance of Evolutionary Optimization by. Dynamically Scaling the Evaluation Function.

Igor A. Khovanov,Vadim S. Anishchenko, Astrakhanskaya str. 83, Saratov, Russia

1/sqrt(B) convergence 1/B convergence B

Repeated Occurrences of the Baldwin Effect Can Guide Evolution on Rugged Fitness Landscapes

A comparison of evolutionary and coevolutionary search

Stochastic Learning in a Neural Network with Adapting. Synapses. Istituto Nazionale di Fisica Nucleare, Sezione di Bari

Learning Conditional Probabilities from Incomplete Data: An Experimental Comparison Marco Ramoni Knowledge Media Institute Paola Sebastiani Statistics

Introduction to Walsh Analysis

SI Appendix. 1. A detailed description of the five model systems

Carnegie Mellon University Forbes Ave. Pittsburgh, PA 15213, USA. fmunos, leemon, V (x)ln + max. cost functional [3].

Metastable Evolutionary Dynamics: Crossing Fitness Barriers or Escaping via Neutral Paths?

Lecture 2: Review of Prerequisites. Table of contents

(Anti-)Stable Points and the Dynamics of Extended Systems

Unstable K=K T=T. Unstable K=K T=T

6. APPLICATION TO THE TRAVELING SALESMAN PROBLEM

V. Evolutionary Computing. Read Flake, ch. 20. Genetic Algorithms. Part 5A: Genetic Algorithms 4/10/17. A. Genetic Algorithms

Schema Theory. David White. Wesleyan University. November 30, 2009

The Evolution of Emergent Computation

1. Computação Evolutiva

Linearly-solvable Markov decision problems

Dynamical Systems. August 13, 2013

Chaos and Liapunov exponents

Stochastic Dynamics of Learning with Momentum. in Neural Networks. Department of Medical Physics and Biophysics,

Optimum Signature Sequence Sets for. Asynchronous CDMA Systems. Abstract

arxiv: v1 [q-bio.pe] 28 Apr 2007

Fitness distributions and GA hardness

Genetic Algorithms and Genetic Programming Lecture 17

2 IGOR PAK so we loose some information about the structure of the tilings since there could be many tilings of with the same multiset of tiles (see e

Evolving Globally Synchronized Cellular Automata

Development. biologically-inspired computing. lecture 16. Informatics luis rocha x x x. Syntactic Operations. biologically Inspired computing

V. Evolutionary Computing. Read Flake, ch. 20. Assumptions. Genetic Algorithms. Fitness-Biased Selection. Outline of Simplified GA

ARTIFICIAL INTELLIGENCE LABORATORY. and CENTER FOR BIOLOGICAL INFORMATION PROCESSING. A.I. Memo No August Federico Girosi.

Analyzing the Evolutionary Pressures in XCS Martin V. Butz Institute for Psychology III University of Wurzburg Wurzburg, 977, Germany

Leslie Matrix II Formal Demography Stanford Spring Workshop in Formal Demography May 2008

Field indeced pattern simulation and spinodal. point in nematic liquid crystals. Chun Zheng Frank Lonberg Robert B. Meyer. June 18, 1995.

State of the art in genetic algorithms' research

DRAFT -- DRAFT -- DRAFT -- DRAFT -- DRAFT --

Linear Algebra of the Quasispecies Model

Linear Regression and Its Applications

IV. Evolutionary Computing. Read Flake, ch. 20. Assumptions. Genetic Algorithms. Fitness-Biased Selection. Outline of Simplified GA

CONTROL SYSTEMS, ROBOTICS AND AUTOMATION Vol. XI Stochastic Stability - H.J. Kushner

Lyapunov exponents in random Boolean networks

GENETIC ALGORITHM FOR CELL DESIGN UNDER SINGLE AND MULTIPLE PERIODS

An Analysis of Diploidy and Dominance in Genetic Algorithms

Comparison of Evolutionary Methods for. Smoother Evolution. 2-4 Hikaridai, Seika-cho. Soraku-gun, Kyoto , JAPAN

The Evolution of Gene Dominance through the. Baldwin Effect

Transcription:

Finite Populations Induce Metastability in Evolutionary Search Erik van Nimwegen y James P. Crutcheld, yz Melanie Mitchell y y Santa Fe Institute, 99 Hyde Park Road, Santa Fe, NM 8750 z Physics Department, University of California, Berkeley, CA 9470-700 (January, 997) Santa Fe Institute Working Paper 96-08-054 to appear in Physics Letters A (997) We introduce an analytical model that predicts the dynamics of a simple evolutionary algorithm in terms of the ow in the space of tness distributions. In the limit of innite populations the dynamics is derived in closed form. We show how nite populations induce periods of stasis \tness epochs" and rapid jumps \innovations". The analysis identies the epochs with the ow's metastable xed points and gives exact predictions of epoch tness level and population distribution. Evolutionary search comprises a set of optimization and learning techniques based loosely on ideas from biological evolution in which individuals in a population (typically candidate solutions to a given problem) are selected for replication based on a \tness" measure dened by some externally imposed criterion. Genetic variation is introduced into the population by stochastically modifying the selected individuals. In practice evolutionary search methods have often been applied successfully to optimization and learning problems in in poorly understood high-dimensional problem spaces. For overviews of such methods and their applications, see, e.g., []. In the following we analyze one particular evolutionary search method: a simplied genetic algorithm (GA). One common feature of the dynamics of evolutionary search is metastable behavior the average tness in the population goes through periods of stasis, which we will call \tness epochs", alternating with rapid jumps in average tness, which we will call \innovations". Such behavior has been reported in many applications of evolutionary search as well as in models of evolution (e.g., []). As has often been found, the length, location, and shape of these tness epochs and innovations depend on dierent control parameters of the evolutionary method, such as population size and mutation and crossover rates. This dependence is complicated and, as yet, poorly understood. There has been a moderate amount of theoretical work on evolutionary algorithms. The approach of Vose, Nix, and Liepins, for example, describes the state of a genetic algorithm at a certain time by avector on the simplex of a high-dimensional Euclidean space. Each dimension of this space represents either a certain genome [] or the state of the population as a whole [4]. The GA dynamics is specied by a nonlinear matrix operator that acts on this vector to produce the state at the next time step. Although this formalism exactly captures the detailed \microscopic" GA dynamics, in practice the large size of these matrices makes it impossible to obtain quantitative results. Here, a matrix operator will be constructed that is similar in spirit to the Vose et al. operators but that has vastly reduced dimensionality since it acts on vectors representing tness distributions. This makes a quantitative analysis of this operator possible, allowing specic quantitative predictions to be made about the GA's behavior. Another theoretical approach is a formalism developed by Prugel-Bennett, Shapiro, and Rattray that uses a statistical mechanics approach to analyze the dynamics of genetic algorithms [5]. In their formalism the focus is also on the evolution of tness distributions, but generally only the average evolution of the rst few cumulants of the tness distribution are studied. This averaging of the dynamics over a large number of runs makes it impossible to describe the epoch-innovation structure of the dynamics in which we are interested. The GA's population will consist of bit strings of length L = NK. The tness function f considers each string s to consist of N contiguous blocks of K bits. For each block of length K, f species a particular desired bit conguration (schema). Without loss of generality, this desired conguration can be taken to be K s for each block. The tness f(s) of a string s is then simply the number of blocks in string s that consist of K ones. Thus, 0 f(s) N. This type of tness function, a so-called \Royal Road" function, was initially designed to address questions about the processing and recombination of schemata by GAs. Such block-based tness functions were thought to lay out a \royal road" on which genetic algorithms would optimally search. In fact, the GA dynamics is substantially more complicated than originally hypothesized for these idealized tness functions [6]. Here we use f because it is simple enough to be analyzed and yet the dynamics of the GA on this tness function exhibits the epochal dynamics so often encountered in evolutionary search. Our analysis assumes the following simplied genetic

algorithm. Initially, the GA generates a population P 0 of M bit strings of length L = NK,chosen with uniform probability from the space of all L-bit strings. It then evaluates the tness f(s) of each string s P 0. Next, it creates a new population of M strings, by selecting strings from the current population with replacement, and with probability proportional to tness. Finally, it mutates each selected string at each site with a xed probability q. At this point a new population P has been created and the cycle starts over by re-evaluating the tnesses of the new strings. Formally, the GA behavior is specied by an iterated procedure, P t+ = GP t, in which the dynamic G represents the implementation of the selection and mutation steps. Unlike many GAs this algorithm does not include crossover the exchange of components from two or more parent strings. Since the main qualitative feature of the dynamics the occurrence of tness epochs is not changed by omitting crossover and the analysis is simplied considerably without it, it is not included in this initial work. Also, this GA is simple in that there is a trivial mapping from the genetic specication to a phenotype: the tness is evaluated directly on the genome. <f> 8 7 6 5 4 0 0 500 000 500 000 500 000 500 4000 4500 7.50 7.5 6.75 Generation t 7 6.5 50 00 50 00 50 400 70 80 90 0000040 FIG.. Average tness in the population over 4500 generations for a GA run with N =8, K =8, q =0:00, and M = 500. The horizontal lines are the predicted tness values for the epochs. The inset plots compare the speed of the innovations at early (lower inset) and late (upper inset) generations. At later generations they are slower: note that the upper inset plot covers a time scale that is more than three times longer than the lower. Figures and display the results of two runs of this GA with dierent parameter settings. The average tness in the population hfi over time t (generation number) is shown in both. Figure, in addition, shows the best tness in the population as a function of time. In these plots the tness epochs are clear. The larger block size and the lower mutation rate of the run in Figure seem to lead to more sharply dened epochs than in the run of Figure. We would like tobeabletoexplainwhy the tness follows this pattern of stepwise increase. In particular, a complete analysis of the simplied GA on this class of tness functions would provide an analytical framework that, given a set of parameters (N, K, q, and M), can answer the following questions: (i) How many tness epochs can possibly occur? (ii) At what tness levels do the tness epochs occur? (iii) What is their expected duration? (iv) What is the expected speed of the innovations between tness epochs? In this paper we will answer the rst two questions and give some preliminary insights into the last two. <f> 7 6 5 4 0 0 50 00 50 00 50 00 50 400 450 500 Generation t FIG.. Average tness (solid line) and best tness (diamonds) in the population over 500 generations for a GA run with N =7,K =4,q =0:004, and M = 500. The horizontal lines are the predicted tness values for the epochs. We assume that the dynamics of the population P t can be captured by analyzing the dynamics of the distribution P ~ (t) of tnesses in the population. In the M! limit of innite population size, the dynamics of P ~ (t) will become deterministic and can be solved analytically. A tness distribution P ~ =(P 0 ::: P N ) is a point on the N + dimensional simplex, where P f is the proportion of individuals with tness f, 0 f N, and the distribution is normalized. Thus, the state space is the simplex = f P ~ P R N+ N : f P =0 f =g. The P average N tness hfi in the population is simply hfi = f fp =0 f. To obtain the innite-population dynamics we construct a generation operator G that incorporates the eects of selection and mutation on the population at the level of the time-dependent tness distribution that is, the GA dynamics are nowspeciedby P ~ (t+) = G( P ~ (t)), where G = M S is the product of the mutation and selection operators M and S, respectively. To construct the mutation operator, we assume that all \incorrect" blocks (those not consisting of all s) in the population can be taken to be independent and, furthermore, that all incorrect blocks have not been correct in their past. In other words, strings in the population are not too similar and strings in which blocks have just been destroyed have alow probability of being selected. Under these assumptions, we can consider the dynamics of a string s consisting of a single block ofk bits under mutation alone to obtain an approximation to the probability that mutation will turn an incorrect block into a correct block. At each time step, each bit in s is mu-

tated with probability q. Starting from a random K-bit string at time t =0,a Markov chain analysis gives the average time T (q K) until the correct constellation of K s appears for the rst time. Our calculation shows that the rst-passage-time distribution for this problem is exponential to a high approximation. As a result, the probability A that an \incorrect" block will be transformed into a correct block in one time step is constant: A ==T (q K). The probability D that a correct block is transformed into an incorrect block in one time step is D =; ( ; q) K. Using the single-block expressions for A and D we can determine the probability M ij that a string s of length NK with tness j will be transformed into a string with tness i in one time step. M ij is the sum over all the probabilities that k incorrect blocks in the string will become correct in that time step and l correct blocks will become incorrect in that time step, such that j+k;l = i. The matrix entries of the mutation operator M are thus given by P N;j P M ij = j ; N;j k=0 l=0 j+k i+l k A k ( ; A) N;j;k D l ( ; D) j;l () ; j l where i j is the Kronecker delta function. To construct the selection operator, recall that our simple GA uses tness-proportionate selection. This means that the fraction Pf s of strings with tness f after selection is proportional to both f and the fraction P f of strings with tness f before selection. That is, Pf s / fp f. Since P ~ s remains normalized, it can easily be seen that Pf s = fp f =hfi, wherehfi is the mean tness of the distribution P ~. We can therefore write the matrix entries S ij of the selection operator S as S ij = i ij =hfi. Notice that, in contrast to the mutation operator M, the selection operator S is nonlinear: S scales with the inverse of the average tness hfi. The generation operator G is simply the product of mutation and selection operators M and S: G ij = NX k=0 M ik S kj : () Note that the tness distribution in the population at time t is given by thet'th iterate of G: ~P (t) =G t [ ~ P (0)]: () To analyze G t we will rst construct a linearized version ~G = M ~S, linearizing the selection operator by setting S = ~S=hfi. Since ~G is proportional to G at each time step, we can formally write the t'th power of G as: G t [ ~ P (0)] = C[t ~ P (0)] ~G t ~ P (0) (4) where C[t ~ P (0)] is a constant that depends on t and the tness distribution at time t =0. Since, in general, ~G has N + distinct eigenvalues f i and N + distinct normalized eigenvectors V ~ i, its diagonal form is given by R ; ~G R, where R ij = V j i is the similarity transformation matrix. The tness distribution in this basis is ~P (t) = NX i=0 i (t) ~ V i : (5) and we canthink of a given ~ P (t) as a distribution ~(t) over the dierent eigenvectors ~V i. In this basis, equation takes on a very simple form: i (t) =C[t ~(0)] f t i i (0) = f t i i(0) Pi f t i i(0) (6) where the constant C[t ~(0)] is easily determined by demanding that the distribution ~(t) over the eigenvectors be normalized. Equation 6 shows that the fractions i (t) of the nonprincipal eigenvectors V ~ i eventually all become exponentially damped. In the limit of t!the fraction N of the principal eigenvector V ~ N exponentially approaches, [7]. Furthermore, we obtain a simple expression for the average tness in the population as a function of time: Pi hf(t)i = P fi t+ i (0) i f i t i(0) : (7) Figure shows the predicted innite-population dynamics for N =,K =4andq =0:0 together with the average experimental results over 0 runs of the GA using those parameters and a large population (M =0 4 ). The error-bars denote plus or minus two standard deviations from the average over the 0 runs. For large populations (M > NK ) the actual dynamics is accurately predicted by the innite-population dynamics. Notice that for populations of this size the average tness increases smoothly as a function of time and there are no discernible tness epochs. <f>.5.5 0.5 0 0 5 0 5 0 5 0 5 40 Generation t FIG.. Theoretical prediction of average tness over time for the parameters N =, K =4, q =0:0, and M =0 4 (solid line) along with average experimental results over 0 runs with the same parameters. The error-bars give plus and minus two standard deviations from the experimental average.

Several additional properties of the GA's behavior follow directly from our analysis of the innite-population dynamics given in equation 7. First, for any >0 and 0 <<andany nite number of time steps t, thereis a population size M such that with probability greater than the nite-population trajectory stays within a distance of the innite-population trajectory ~ P (t) for all t time steps, for population sizes M>M. (The proof of this is analogous to a similar proof in [4].) Second, the average tness hf(t)i is a monotonically increasing function of time. Finally, using Perron's theorem for positive matrices, wecanshow that the innite-population dynamics has a unique asymptotic xed-point tness distribution ~P () which is given by ~G's principal eigenvector ~ V N. For nite populations of size M, the tnessdistribution state space is a lattice M over the N + dimensional simplex with lattice spacing M ; : M = f ~ P : P f = n f M n f N NX f =0 n f = Mg (8) where n f is the number of individuals with tness f in the population. As in the innite-population case, applying G to a tness distribution ~ P in M gives the expected tness distribution h ~ P 0 i = G( ~ P ) at the next generation. The actual tness distribution ~ P 0 at the next time step will be a random sample of size M of the distribution G( ~ P ). This induces stochasticity in the nite-population dynamics. In general, for a nite population of size M, the probability Pr[ ~ P! ~ P 0 ] that a tness distribution P f = n f =M will go to a tness distribution P 0 f = m f =M under mutation and selection will be given by a multinomial distribution with mean G( ~ P ). That is, Pr[ ~ P! ~ P 0 ]=M! NY f =0 [G f ( ~ P )] m f m f! (9) where G f ( P ~ ) is the expected proportion hpf 0 i of individuals with tness f at the next generation. The expected change hp f i in the fth component of the tness distribution is given by the dierence G f ( P ~ ) ; P f. If hp f i is much smaller than the lattice spacing =M, thenumber of individuals with tness f most likely will not change. Imagine the case where no individual has a tness higher than k, i.e. P f =0 f >k, and the population is conned to the k + dimensional subsimplex k M. If the probability of creating new blocks is small with respect to =M then the higher components (f > k) of the tness distribution are likely to remain 0 for some time and the population will stay in k M. In other words, the ow hp f i out of the subsimplex (f >k) is small with respect to the lattice spacing =M. At the same time, mutation (which continues to destroy blocks) and selection will steer the population to a metastable xed point within k M, where it remains until anewblock is created and the associated string spreads in the population. This metastable xed point isgiven by the largest eigenvalue of the k + -dimensional submatrix of ~G. For small mutation rates q it can be shown that the largest eigenvectors of the k + -dimensional submatrices lie close to the eigenvectors V ~ k of the full matrix ~G. So, although the generation operator G has only one xed point in, it's nonprincipal eigenvectors V ~ k lie very close to the ow's metastable xed points in the lower dimensional subsimplices k M, around which the population stabilizes until strings with tness f > kare discovered. In short, the combination of the state-space discreteness and the stochasticity of the GA dynamics both induced by the population's nite size causes populations moving in k M to temporarily stabilize within these subsimplices. In this way, metastability is induced via nite populations. Using the fact that the mutation operator M is a stochastic matrix, we can show that each eigenvalue f i of ~G is equal to the average tness hfi in the population for the tness distributions of the corresponding eigenvector ~V i. We therefore expect the tness epochs to occur at tness levels that simply correspond to the eigenvalues of ~G. Using a second order perturbation analysis in q of the eigenvalues of ~G, the tness levels are approximated, at low mutation, by f i (q) = i ; [i K + i(n ; i)a ]q + O(q ) (0) where A q is the probability to align a block to rst order in P q that is, A = A q + O(q ) and where A ; = ;K K j;; K ; Pj; ; K. j= j i=0 i In addition to the experimental data from the GA simulation runs, Figures and show the tness values corresponding to the eigenvalues f i of ~G, which were calculated numerically, as horizontal dashed lines. It is clear that the eigenvalues accurately predict the tness levels at which theepochs occur. Figures and show that epoch duration varies among the dierent epochs and that not all possible tness epochs occur in any particular run of the genetic algorithm. Moreover, we observe thatonaverage the higher tness epochs have longer durations and are more likely to be visited in a given run. These observations can be understood by a linear stability analysis of G around the metastable states V ~ k. Solving for the spectrum of the Jacobian matrix DG, we nd that the eigenvalues k i of the Jacobian matrix at V ~ k are given by: k i = f i(q) (k 6= i) () f k (q) From this, we see that the number of stable dimensions of a metastable state ~ V k is equal to k (where 0 k N). Since the actual center of the tness epoch, given by the largest eigenvalue of the k + -dimensional submatrix of 4

~G, isslightly oset from V ~ k in the direction of the unstable dimensions, one sees that qualitatively the duration of an epoch is inversely proportional to the number of unstable dimensions and the size of the eigenvalues k i of the unstable dimensions (i > k). For larger k there are fewer unstable dimensions and the sizes of the k i are relatively smaller. Therefore, we expect the higher-tness epochs to have a longer duration on average. The probability that the population will visit a particular epoch is determined by the probability that the population will fall onto one of the attracting dimensions of that epoch in the course of its evolution. Again qualitatively, we expect that an epoch withalargernumber of stable dimensions will therefore also have a higher probability of being visited in any particular run. We see from equations 0 and that for all unstable dimensions ( k i > ) of a metastable state V ~ k we have k i +=k as q! 0. For small values of k (= O()) the eigenvalues corresponding to the unstable dimensions of V ~ k are therefore considerably larger than. This means that as soon as the tness distribution acquires a minimal component of size =M or more in an unstable dimension, it will start moving awayfromthe metastable state exponentially fast. This explains the occurrence of the steep innovations between the epochs. Once an unstable dimension directed towards a higher metastable state is discovered by evolution, the tness distribution will move to this new metastable state exponentially fast. For higher mutation rates, the range of the ratios f i =f k about decreases, and we expect the innovations to become less steep. We indeed see more gradual innovations in the run with the higher mutation rate (Figure ). The eigenvalues f i =f k of the unstable dimensions (i >k) approach from above for increasing values of k. We therefore expect the innovations between higher-tness epochs to be slower than the innovations between lower-tness epochs. This eect is enhanced by the fact that the rst-order term i N increases with i. The inset plots in Figure show that innovations at later generations are indeed slower. The time scale for the upper plot (later generations) is three times that of the lower (earlier generations). Finally, since there are fewer stable dimensions for smaller k and the stable dimension eigenvalues k i are smaller as well, we expect the tness uctuations in the lower-tness epochs to be smaller than the uctuations in the higher-tness epochs. The plots also illustrate this phenomenon. To summarize and conclude: evolutionary search algorithms are stochastic dynamical systems in which a large set of identical elements evolve inparallelandun- der the inuence of one another through a state space. Macroscopic states for these systems are often dened in terms of the rst moments of the distributions over the state variables of the elements. A commonly observed qualitative behavior is that the mean of some state variable alternates between periods of stasis and sudden change. In this paper we analyzed a simplied genetic algorithm as an example of such a system and derived macroscopic equations of motion, from the microscopic dynamics specied by the GA, that explain the occurrence of these periods of stasis and sudden change. We constructed an operator G that describes the GA's deterministic dynamics in the limit of innite populations. By diagonalizing the linearized version ~G we were able to obtain closed-form expressions for the evolution of the population tness distribution. In the nitepopulation case the dynamics was still governed by the operator G. Due to the nite population, however, the state space became discrete and the nite-population sampling noise caused the dynamics to become stochastic. This, in turn, led to the appearance of metastable tness distributions in the vicinityofthehyperbolic xed points of G. We were able to identify the tness epochs in the dynamics with these hyperbolic xed points. This gave us both the tness levels at which the epochs occur and the associated metastable tness distributions. We also explained several other dynamical features, such as the innovations' short durations and the appearance and disappearance of epochs. The behavior of evolutionary search algorithms is often informally described as moving along a \tness landscape" directly dened by a tness function cf. []. It is clear from our experiments and our analysis of the underlying mechanisms that this geographic metaphor is misleading. First, the tness function is only a partial determinant of the search dynamics. Population size, mutation rate, and other parameters can radically alter the time-dependent metastable behavior revealing, hiding, or even inducing much of the structure in the tness landscape. Note, for instance, that our tness function is simple: it has a single peak, but otherwise no local optima, and it is not \rugged". Second, our analysis suggests that by combining the tness landscape with the population dynamics as we did in constructing G one can obtain an eective, but dierent, \landscape" that guides the evolutionary search. This work was supported at the Santa Fe Institute by NSF IRI-9000, DOE DE-FG0-94ER5, and ONR N0004-95--0975, and at UC Berkeley under ONR N0004-95--054. [] T. Back, Evolutionary Algorithms in Theory and Practice: Evolution Strategies, Evolutionary Programming, Genetic Algorithms, Oxford University Press, Oxford (996) D. B. Fogel, Evolutionary Computation: Toward a New Philosophy of Machine Intelligence, IEEE Press, New York (995) D. E. Goldberg, Genetic Algorithms in Search, Optimization, and Machine Learning, Addison-Wesley, Reading, MA (989) Z. Michalewicz, Genetic Algorithms + Data Structures = Evolution Programs, Articial Intelli- 5

gence Series. Springer-Verlag (99) M. Mitchell, An Introduction to Genetic Algorithms, MIT Press, Cambridge, MA (996). [] S. Wright, Evolution 6 (98) 47-44 A. Bergman and M. W. Feldman, \Question marks about the period of punctuation", Santa Fe Institute Working Paper 96-0- 006 (996) M. Huynen, P. F. Stadler, and W. Fontana, Proc. Natl. Acad. Sci. 9 (996) 97-40 K. Lindgren, in C. G. Langton, C. Taylor, J. D. Farmer, and S. Rasmussen, editors, Articial Life II, Addison-Wesley, Reading, MA (99) 95- J. P. Crutcheld and M. Mitchell, Proc. Natl. Acad. Sci. 9 (995) 074-0746 M. Mitchell, J. P. Crutcheld and P. T. Hraber, Physica D 75 (994) 6-9 T. S. Ray, in C. G. Langton, C. Taylor,J.D.Farmer, and S. Rasmussen, editors, Articial Life II, Addison- Wesley, Reading, MA (99) 7-408. [] M. D. Vose and G. E. Liepins, Complex Systems 5 (99) -44. [4] A. E. Nix and M. D. Vose, Ann. Math. Art. Intell. 5 (99) 79-88. [5] A. Prugel-Bennett and J. L. Shapiro, Phys. Rev. Lett. 7 (994) 05-09 A. Prugel-Bennett and J. L. Shapiro, Physica D (996) in press M. Rattray and J. L. Shapiro, J. of Phys. A 9 (996) 745-747. [6] S. Forrest and M. Mitchell, in L. D. Whitley, editor, Foundations of Genetic Algorithms, Morgan Kauman, San Mateo, CA (99) 09-6 M. Mitchell, J. H. Holland, and S. Forrest, in Advances in Neural Information Processing Systems 6, Morgan Kaufmann, San Mateo, CA (994) 5-58. [7] Our equation, for innite populations, is analogous to a discrete-time version of the dierential equations of the Eigen model of molecular evolution: M. Eigen, Naturwissenschaften 58 (97) 465-5 M. Eigen and P. Schuster, Naturwissenschaften 64 (977) 54-565. The asymptotically dominant eigenvector V ~ N of our model is analogous to a so-called \quasi-species" of molecules: an asymptotically stable distribution over dierent molecular species. Our analysis shows that the evolutionary dynamics for - nite populations can get trapped in regions close to the nonprincipal eigenvectors V ~ i of the generation operator. This suggests that in practical instances of molecular evolution, where the population size is typically much smaller than the size of the sequence space, \metastable quasispecies" are to be expected in the vicinity of the nonprincipal eigenvectors V ~ i of the generation operator (or transfer-matrix, as it has been called in reference to the Eigen model: I. Leuthausser, J. of Stat. Phys. 48 (987) 4-60). 6