Looking Under the EA Hood with Price s Equation

Similar documents
Applying Price s Equation to Survival Selection

Exploration of population fixed-points versus mutation rates for functions of unitation

Polynomial Approximation of Survival Probabilities Under Multi-point Crossover

Representation and Hidden Bias II: Eliminating Defining Length Bias in Genetic Search via Shuffle Crossover

Analyzing Cooperative Coevolution with Evolutionary Game Theory

Evolution Strategies for Constants Optimization in Genetic Programming

Introduction to Walsh Analysis

EClab 2002 Summer Lecture Series

Search. Search is a key component of intelligent problem solving. Get closer to the goal if time is not enough

An Effective Chromosome Representation for Evolving Flexible Job Shop Schedules

On the Strength of Size Limits in Linear Genetic Programming

INVARIANT SUBSETS OF THE SEARCH SPACE AND THE UNIVERSALITY OF A GENERALIZED GENETIC ALGORITHM

Scaling Up. So far, we have considered methods that systematically explore the full search space, possibly using principled pruning (A* etc.).

Genetic Algorithm for Solving the Economic Load Dispatch

Scalability of Selectorecombinative Genetic Algorithms for Problems with Tight Linkage

Running time analysis of a multi-objective evolutionary algorithm on a simple discrete optimization problem

1. Computação Evolutiva

Program Evolution by Integrating EDP and GP

Crossover, Sampling, Bloat and the Harmful Effects of Size Limits

Geometric Semantic Genetic Programming (GSGP): theory-laden design of semantic mutation operators

INFORMATION-THEORETIC BOUNDS OF EVOLUTIONARY PROCESSES MODELED AS A PROTEIN COMMUNICATION SYSTEM. Liuling Gong, Nidhal Bouaynaya and Dan Schonfeld

Bistability of the Needle Function in the Presence of Truncation Selection

New Epistasis Measures for Detecting Independently Optimizable Partitions of Variables

Gaussian EDA and Truncation Selection: Setting Limits for Sustainable Progress

Forecasting & Futurism

Chapter 8: Introduction to Evolutionary Computation

Evolutionary Computation

Fundamentals of Genetic Algorithms

Predicting Dynamical Systems

Implicit Formae in Genetic Algorithms

Artificial Intelligence Methods (G5BAIM) - Examination

Schema Theory. David White. Wesleyan University. November 30, 2009

Crossover Gene Selection by Spatial Location

Fitness distributions and GA hardness

Bounding the Population Size in XCS to Ensure Reproductive Opportunities

Evolutionary Algorithms

Repeated Occurrences of the Baldwin Effect Can Guide Evolution on Rugged Fitness Landscapes

Introduction to Evolutionary Computation

A Tractable Walsh Analysis of SAT and its Implications for Genetic Algorithms

Crossover Techniques in GAs

Data Warehousing & Data Mining

Genetic Engineering and Creative Design

Mating Restriction and Niching Pressure: Results from Agents and Implications for General EC

On Identifying Global Optima in Cooperative Coevolution

Lecture 9 Evolutionary Computation: Genetic algorithms

LandscapeEC: Adding Geographical Structure to Cellular Evolutionary Algorithms

Expected Running Time Analysis of a Multiobjective Evolutionary Algorithm on Pseudo-boolean Functions

Evolutionary computation

Artificial Intelligence (AI) Common AI Methods. Training. Signals to Perceptrons. Artificial Neural Networks (ANN) Artificial Intelligence

Evolutionary Computation: introduction

A Statistical Genetic Algorithm

Gene Pool Recombination in Genetic Algorithms

Investigation of Mutation Strategies in Differential Evolution for Solving Global Optimization Problems

DESIGN OF OPTIMUM CROSS-SECTIONS FOR LOAD-CARRYING MEMBERS USING MULTI-OBJECTIVE EVOLUTIONARY ALGORITHMS

The Role of Crossover in Genetic Algorithms to Solve Optimization of a Function Problem Falih Hassan

Local Search & Optimization

A Note on Crossover with Interval Representations

Three Steps toward Tuning the Coordinate Systems in Nature-Inspired Optimization Algorithms

Quantitative trait evolution with mutations of large effect

Rigorous Analyses for the Combination of Ant Colony Optimization and Local Search

Matrix Analysis of Genetic Programming Mutation

Evolutionary Design I

Evolutionary Theory. Sinauer Associates, Inc. Publishers Sunderland, Massachusetts U.S.A.

Using schema theory to explore interactions of multiple operators

Three Steps toward Tuning the Coordinate Systems in Nature-Inspired Optimization Algorithms

Lecture 13: Simple Linear Regression in Matrix Format

AN EVOLUTIONARY ALGORITHM TO ESTIMATE UNKNOWN HEAT FLUX IN A ONE- DIMENSIONAL INVERSE HEAT CONDUCTION PROBLEM

Time Series 2. Robert Almgren. Sept. 21, 2009

Co-evolutionary dynamics on a deformable landscape

Lecture 22. Introduction to Genetic Algorithms

A schema theory analysis of mutation size biases in genetic programming with linear representations

Multi-objective approaches in a single-objective optimization environment

Background literature. Data Mining. Data mining: what is it?

Stochastic Search: Part 2. Genetic Algorithms. Vincent A. Cicirello. Robotics Institute. Carnegie Mellon University

Multiobjective Optimization of an Extremal Evolution Model

Discrete evaluation and the particle swarm algorithm

Truncation Selection and Gaussian EDA: Bounds for Sustainable Progress in High-Dimensional Spaces

Introduction to Optimization

An Evolution Strategy for the Induction of Fuzzy Finite-state Automata

2 Differential Evolution and its Control Parameters

REVIEW 8/2/2017 陈芳华东师大英语系

Decomposition and Metaoptimization of Mutation Operator in Differential Evolution

Threshold Selection, Hypothesis Tests, and DOE Methods

Evolving Presentations of Genetic Information: Motivation, Methods, and Analysis

When to Use Bit-Wise Neutrality

Local Search & Optimization

Pure Strategy or Mixed Strategy?

Wavelet Transform And Principal Component Analysis Based Feature Extraction

Title: On Finding Effective Courses of Action in a Complex Situation Using Evolutionary Algorithms *

the yellow gene from each of the two parents he wrote Experiments in Plant

N/4 + N/2 + N = 2N 2.

Evolutionary Computation. DEIS-Cesena Alma Mater Studiorum Università di Bologna Cesena (Italia)

Probability (Phi) Proportion BBs (Pbb) 1. Proportion BBs (Pbb) Pop Size (n) Pop Size (n)

Haploid & diploid recombination and their evolutionary impact

UNIVERSITY OF DORTMUND

14 Branching processes

Using phylogenetics to estimate species divergence times... Basics and basic issues for Bayesian inference of divergence times (plus some digression)

Improving Coordination via Emergent Communication in Cooperative Multiagent Systems: A Genetic Network Programming Approach

6 Introduction to Population Genetics

The Evolution of Gene Dominance through the. Baldwin Effect

Transcription:

Looking Under the EA Hood with Price s Equation Jeffrey K. Bassett 1, Mitchell A. Potter 2, and Kenneth A. De Jong 1 1 George Mason University, Fairfax, VA 22030 {jbassett, kdejong}@cs.gmu.edu 2 Naval Research Laboratory, Washington, DC 20375 mpotter@aic.nrl.navy.mil Abstract. In this paper we show how tools based on extensions of Price s equation allow us to look inside production-level EAs to see how selection, representation, and reproductive operators interact with each other, and how these interactions affect EA performance. With such tools it is possible to understand at a deeper level how existing EAs work as well as provide support for making better design decisions involving new EC applications. 1 Introduction Evolutionary algorithm design is difficult for a number of reasons, not the least of which is that the choices of selection, representation, and reproductive operators interact in non-linear ways to affect EA performance. As a consequence, EA designers are often faced with their own black box" optimization problem in terms of finding combinations of design choices that improve EA performance on their particular application domain. Results from the EC theory community continue to provide new insights into this difficult design process, but often are obtained by make simplifying assumptions in order to make the mathematics tractable. This often leaves open the question as to whether particular theoretical results apply in practice to actual EAs being used and/or to newly designed EAs. The results presented here are part of an ongoing effort that tries to bridge this gap by using theoretical results to help us build useful tools that can be used to look inside actual EAs in order to better understand what s happening under the hood". In particular, we have been exploring the use of some theoretical work done by Price (1970) to obtain deeper insights into the interactions of selection, representation, and the reproductive operators of crossover and mutation. In this respect we are indebted to Lee Altenberg who for some time now has been encouraging the EC community to pay more attention to Price s Theorem (Altenberg 1994; Altenberg 1995). In section 2 we provide a brief overview of Price s Theorem and how it can be extended in a way to provide a useful analysis tool. Section 3 describes how Price s Theorem can be further extended to provide additional insights. We illustrate these ideas in section 4 by using the developed tools to instrument actual production-level EAs and showing how two EAs with similar "black box" behavior look quite different under the hood". Finally, in section 5 we conclude with some observations and future work. K. Deb et al. (Eds.): GECCO 2004, LNCS 3102, pp. 914 922, 2004. c Springer-Verlag Berlin Heidelberg 2004

Looking Under the EA Hood with Price s Equation 915 2 Background In 1970, George Price published the article Covariance and (Price 1970) in which he presented an equation that has proved to be a major contribution to the field of evolutionary genetics (Frank 1995). The equation describes the existence of a covariance relationship between the number of successful offspring that an individual produces and the frequency of any given gene in that individual. If this covariance value is high, then the existence of that gene is a good predictor of selection. 2.1 Price s Equation Although Price focused on gene frequency, his equation is more general and can be used to estimate the change in any measurable attribute from the parent population to the child population, and separate the change attributable to selection from the change attributable to the genetic operators. Specifically, Q = Cov(z,q) zi q i +, (1) z where q i is the measurement of some attribute of parent i such as the number of occurrences of a particular gene or combination thereof, z i is the number of children to which parent i contributed genetic material, z is the average number of children produced by each parent, N is the number of parents, and q i is the difference between the average q value of the children of i and the q value of parent i. Price s Equation combines the effect of all the genetic operators into a single term. To further separate the effects of the individual reproductive operators, Potter et al. (2003) extended the equation as follows: Q = Cov(z,q) z + P zi q ij, (2) where P is the number of genetic operators and q ij is the difference between the average q value of the children of i measured before and after the application of operator j. j=1 2.2 Previous Applications in Evolutionary Computation Altenberg (1995) was one of the first in the EC community to call attention to Price s Equation. He demonstrated Price s assertion that gene frequency is not the only attribute of the individuals which can be predicted by the equation, and identified several different measurement functions which could be useful, including mean fitness from both the biological and evolutionary computation perspectives, frequency of schemata, and evolvability. More recently, Langdon and Poli (2002) showed how measuring gene frequencies is equivalent to determining the frequency of use of the available primitives in the evolving solution trees, and used Price s Equation to diagnose the probable causes of poorer performing runs.

916 J. Bassett, M.A. Potter, and K.A. De Jong Finally, Potter, Bassett, and De Jong (2003) concentrated on using fitness as a measurement function and applied Price s Equation to the visualization of the dynamics of evolvability, that is, the ability of an EA to continue to make improvements in fitness over time. 3 Variance of Operator Effects Each term in Price s equation calculates the contribution of a different operator as a mean of the attribute being measured. Although visualizations based on the decomposition of delta mean fitness into operator-based components certainly provides more information than simple best-so-far curves (Potter et al. 2003), focusing on the mean can sometimes be misleading. In particular, the mean may be close to zero, leading one to believe that an operator is making a minimal contribution, when in fact it is a critical component of the evolutionary process. In fact, the average individuals are not the ones driving evolution forward. It is the occasional exceptional individuals created by crossover and mutation that enable the population to continue to improve over time. The best and worst individuals are at the upper and lower tails of the population fitness distributions. Therefore, if we want to get some sense of how often an operator creates above or below average individuals, we need to look at the variance of the q values for mutation and crossover, not just the mean. We should emphasize that we are interested in the variance of the effect of an operator on the individuals of a population, not the variance of the mean effect given multiple runs. Specifically, let E[X] be the expected change in the measurement of q due to a particular operator such as crossover. For simplicity we will assume for the moment that only a single operator is used. Expanding the second term of equation 1 we have N i=1 E[X] = z i q i zi N i=1 = z k=1 (q ik q i) i z i N zi i=1 k=1 = (q ik q i ), where q ik is the measured q of the kth child of parent i, N is the number of parents, z i is the number of children produced by parent i, and z is the average number of children produced by each parent. From this expansion we see that the random variable of interest is X = q ik q i, and N ( zi i=1 k=1 Var[X]= (q N ik q i ) 2 ) zi 2 i=1 k=1 (q ik q i ). (3) This can be extended to multiple operators by expanding equation 2, resulting in X = q ijk q i(j 1)k, where q ijk is the measured q of the kth child of parent i after the application of operator j.

4 Looking Under the Hood Looking Under the EA Hood with Price s Equation 917 To illustrate how to use these ideas to look under the hood, we will compare the performance of our evolutionary algorithm using two different mutation operators. In particular we will focus on the importance of the fitness variance in relation to Price s Equation. 4.1 Example Problem To help in our illustration, we have chosen a standard problem from the function optimization literature introduced by Schwefel (1981). The objective function f(x) = 418.9829n + n i=1 ) x i sin( xi defines a landscape covered with a lattice of large peaks and basins. The predominant characteristic of the Schwefel function is the presence of a second-best maximum far away from the global maximum, intended to trap optimization algorithms on a suboptimal peak. The best maximums are near the corners of the space. In this formulation of the problem, the global minimum is zero and the global maximum is 837.9658n. In our experiments the problem has thirty independent variables constrained to the range ( 500.0, 500.0). 4.2 Algorithm Design Typically as one makes design decisions for an evolutionary algorithm they go through a process of trial and error. Most of us simply exchange components, perform some runs, and then compare the best-so-far curves. We find a component that seems to work well for our problem, and then move on to other design decisions or parameter tuning, hoping that all of our choices will complement each other. Let us assume that we have already made a series of design decisions, including using a real-valued representation and (µ, λ) selection of (500, 1000). We ve chosen these unusually large population sizes in order to increase the sample sizes and reduce the amount of noise when calculating the terms in Price s equation. We ve also decided to use two-point crossover at a rate of 0.6 and we ve implemented it all with the ECKit Java class library developed by Potter (1998). Now we are left with one final decision, the mutation operator. We are going to use a Gaussian mutation which is applied to all genes in the genome, but we want to decide between using a fixed standard deviation or one where the standard deviations are adapted, as described in (Bäck and Schwefel 1993). 4.3 Comparing Operators In order to compare operators, we perform 100 runs using the fixed Gaussian mutation (with a standard deviation of 1.0), and 100 runs with the adaptive gaussian mutation. The best-so-far curves for each are plotted Figure 1. Based on what we see here, both

918 J. Bassett, M.A. Potter, and K.A. De Jong Fitness 16000 20000 24000 Fixed mutation Adaptive mutation 0 200 400 600 800 200 Fig. 1. Comparison of best fitness curves from Schwefel function optimization using fixed and adaptive Gaussian mutations. The global optimum is approximately 25,140. Fig. 2. The mean contributions from selection, crossover and mutation during optimization of the Schwefel function using fixed Gaussian mutation. These were calculated using the extended Price s equation. appear to be able to reach the optimum consistently, but the fixed Gaussian mutation operator seems to be able to do so more quickly. Without any other information, this is the operator we would choose. But do we have any idea why this operator is better than the other? The EA is essentially a black box, and we have very little idea of what is happenning inside. If we could get more information about how the operators are interacting and what effects they are having, we could make more intelligent decisions about how to improve the performance of our EA. By applying the extended version of Price s Equation (equation 2) to the EA while it is running, we can calculate the average effect that each operator has on the population. Figure 2 shows the mean effects of selection, crossover and mutation when using the fixed Gaussian mutation operator. The plot shows that the crossover and mutation operators have an average effect near zero (which may be difficult to read in the plot), while selection seems to be doing all of the work. What exactly does this mean? These average values are clearly not telling us the whole story. Recall that in section 3 we claimed that the average individuals created by the genetic operators are not as important as the exceptional ones because it is the exceptional individuals which are chosen by selection. We want to look at the standard deviations of the q values, but instead of plotting them in a separate plot, we have combined it with the existing plots of the delta mean fitnesses. Figure 3 gives an example of this type of plot for the crossover operator. For each generation a gray band is drawn starting from the mean effect of crossover (which is very close to zero) up to plus one standard deviation and down to minus one standard deviation. This of course assumes that the fitness distribution of the q values is normal, which is not always a good assumption.

Looking Under the EA Hood with Price s Equation 919 600 200 0 200 400 600 600 200 0 200 400 600 Fig. 3. Standard deviation of delta fitness effects caused by the crossover operator. The experiment was run using a fixed Gaussian mutation with standard deviation of 1.0 on the Schwefel function. Results are averaged over 100 runs. Fig. 4. Standard deviation of delta fitness effects caused by the mutation operator. The experiment was run using a fixed Gaussian mutation with standard deviation of 1.0 on the Schwefel function. Results are averaged over 100 runs. The advantage of this type of plot is that one can compare the relative effects of the delta mean fitness to the effects of the variance. In other words, an operator may on average be quite disruptive (low mean fitness), but still have a high variance. This would mean that the operator could still be effective. One should keep in mind though that we are plotting only one standard deviation. Genetic operators can often create individuals with fitnesses that are two to three standard deviations from the mean when run on this problem. Getting back to our example, Figures 3 and 4 show the variance effects of crossover and fixed Gaussian mutation. Whereas before the two operators were indistinguishable, now we can see a large difference between them. Notice that the upper edge of the crossover standard deviation curve follows the curve representing mean contribution from selection very closely. It seems clear that crossover is contributing more to the search process than mutation, at least in the early generations. The variance of the mutation operator is much lower, as can be seen by the very narrow gray band along the x-axis. It is unlikely mutation does much more than local hill climbing. Moving on to the adaptive Gaussian mutation operator, Figure 5 shows the mean effects of selection, crossover and mutation when using this operator. Here we see that once again the average effect of crossover is close to zero, but this time the average effects of mutation are much more negative. The adaptive mutation operator appears to be much more disruptive than the fixed Gaussian mutation. But we should not draw too many conclusions before we see the variance effects of the operators, which are plotted in Figures 6 and 7. One of the first things to note in Figure 7 is that although mutation is very disruptive on average, the upper edge of the standard deviation curve still comes above the x-axis by a fair margin, which indicates

920 J. Bassett, M.A. Potter, and K.A. De Jong 1000 500 0 500 1000 Q Fig. 5. The mean contributions from selection, crossover and mutation during optimization of the Schwefel function using adaptive Gaussian mutation. These were calculated using the extended Price s equation. 2000 1000 0 500 1000 Q 2000 1000 0 500 1000 Q Fig. 6. Standard deviation of delta fitness effects caused by the crossover operator. The experiment was run using an adaptive Gaussian mutation on the Schwefel function. Results are averaged over 100 runs. Fig. 7. Standard deviation of delta fitness effects caused by the mutation operator. The experiment was run using an adaptive Gaussian mutation on the Schwefel function. Results are averaged over 100 runs. that it is contributing to the search process more than the fixed gaussian mutation operator did. There is something more interesting though. Now we can get some insight into why it takes longer to reach the optimum using the adaptive Gaussian mutation than it does using the fixed Gaussian mutation. Compare the crossover standard deviations in Figure 6 with the ones in Figure 3. In conjunction with the adaptive mutation, crossover continues to

Looking Under the EA Hood with Price s Equation 921 have high variances out to generation 50, as opposed to the rapidly decreasing variances we see when fixed Gaussian mutation is used. The disruptive effects of mutation are so high that crossover is having to spend more time repairing individuals. It cannnot make headway in the search process until the standard deviations for mutation have been reduced. With this knowledge we can now make a more informed decision about choosing our mutation operator. It is clear that crossover can take care of exploration on its own (in this domain), so we do not need a mutation operator which performs search also, especially when it comes with the undesirable side effects of disuption. 5 Conclusions and Future Work In this paper we have shown how tools based on extensions of Price s equation allow us to look inside production-level EAs to see how selection, representation, and reproductive operators interact with each other, and how these interactions affect EA performance. In particular, we have shown how these extensions can provide insight into the way in which reproductive operator variance provides the exploratory power needed for good performance. The reported results focused on ES-like EAs. We are in the process of completing a similar study for GA-like EAs. The interesting preliminary results suggest that in this case crossover and mutation interact internally in quite different ways. With results such as these we believe that it is possible to understand at a deeper level how existing EAs work as well as provide support for making better design decisions involving new EC applications. Acknowledgments. We thank Donald Sofge, Magdalena Bugajska, Myriam Abramson, and Paul Wiegand for helpful discussions on the topic of Price s Equation. The work reported in this paper was supported by the Office of Naval Research under work request N0001403WX20212. References [Altenberg 1994] Altenberg, L. (1994). The evolution of evolvability in genetic programming. In K. E. K. Jr. (Ed.), Advances in Genetic Programming, Chapter 3, pp. 47 74. MIT Press. [Altenberg 1995] Altenberg, L. (1995). The schema theorem and Price s theorem. In L. D. Whitley and M. D. Vose (Eds.), Foundations of Genetic Algorithms III, pp. 23 49. Morgan Kaufmann. [Bäck and Schwefel 1993] Bäck, T. and H.-P. Schwefel (1993). An overview of evolutionary algorithms for parameter optimization. Evolutionary Computation 1(1), 1 23. [Frank 1995] Frank, S. A. (1995). George price s contributions to evolutionary genetics. Journal of Theoretical Biology 175, 373 388.

922 J. Bassett, M.A. Potter, and K.A. De Jong [Langdon and Poli 2002] Langdon, W. B. and R. Poli (2002). Foundations of Genetic Programming. Berlin Heidelberg: Springer-Verlag. [Potter 1998] Potter, M. A. (1998). Overview of the evolutionary computation toolkit. http://cs.gmu.edu/ mpotter/. [Potter, Bassett, and De Jong 2003] Potter, M. A., J. K. Bassett, and K. A. De Jong (2003). Visualizing evolvability with price s equation. In Proceedings of the 2003 Congress on Evolutionary Computation, pp. 2785 2790. IEEE. [Price 1970] Price, G. (1970). and covariance. Nature 227, 520 521. [Schwefel 1981] Schwefel, H.-P. (1981). Numerical optimization of Computer models. Chichester: John Wiley & Sons, Ltd.