Average Drift Analysis and Population Scalability

Similar documents
A Comparison of GAs Penalizing Infeasible Solutions and Repairing Infeasible Solutions on the 0-1 Knapsack Problem

DRAFT -- DRAFT -- DRAFT -- DRAFT -- DRAFT --

Pure Strategy or Mixed Strategy?

A Lower Bound Analysis of Population-based Evolutionary Algorithms for Pseudo-Boolean Functions

Evolutionary Algorithms How to Cope With Plateaus of Constant Fitness and When to Reject Strings of The Same Fitness

When Is an Estimation of Distribution Algorithm Better than an Evolutionary Algorithm?

Crossover can be constructive when computing unique input output sequences

Evolutionary Computation Theory. Jun He School of Computer Science University of Birmingham Web: jxh

When to Use Bit-Wise Neutrality

An Analysis on Recombination in Multi-Objective Evolutionary Optimization

Runtime Analysis of Evolutionary Algorithms for the Knapsack Problem with Favorably Correlated Weights

REIHE COMPUTATIONAL INTELLIGENCE COLLABORATIVE RESEARCH CENTER 531

REIHE COMPUTATIONAL INTELLIGENCE COLLABORATIVE RESEARCH CENTER 531

A New Approach to Estimating the Expected First Hitting Time of Evolutionary Algorithms

Switch Analysis for Running Time Analysis of Evolutionary Algorithms

A Gentle Introduction to the Time Complexity Analysis of Evolutionary Algorithms

Plateaus Can Be Harder in Multi-Objective Optimization

DRAFT -- DRAFT -- DRAFT -- DRAFT -- DRAFT --

Geometric Semantic Genetic Programming (GSGP): theory-laden design of semantic mutation operators

When Hypermutations and Ageing Enable Artificial Immune Systems to Outperform Evolutionary Algorithms 1. Dogan Corus, Pietro S. Oliveto, Donya Yazdani

Geometric Semantic Genetic Programming (GSGP): theory-laden design of variation operators

Black Box Search By Unbiased Variation

On the Impact of Objective Function Transformations on Evolutionary and Black-Box Algorithms

A Runtime Analysis of Parallel Evolutionary Algorithms in Dynamic Optimization

Multiplicative Drift Analysis

How to Treat Evolutionary Algorithms as Ordinary Randomized Algorithms

On the Usefulness of Infeasible Solutions in Evolutionary Search: A Theoretical Study

On the Effectiveness of Sampling for Evolutionary Optimization in Noisy Environments

Black-Box Search by Unbiased Variation

Introduction. Genetic Algorithm Theory. Overview of tutorial. The Simple Genetic Algorithm. Jonathan E. Rowe

REIHE COMPUTATIONAL INTELLIGENCE S O N D E R F O R S C H U N G S B E R E I C H 5 3 1

Runtime Analysis of (1+1) EA on Computing Unique Input Output Sequences

Runtime Analysis of Genetic Algorithms with Very High Selection Pressure

Runtime Analysis of a Binary Particle Swarm Optimizer

Genetic Algorithms: Basic Principles and Applications

Upper Bounds on the Time and Space Complexity of Optimizing Additively Separable Functions

The Fitness Level Method with Tail Bounds

When to use bit-wise neutrality

TECHNISCHE UNIVERSITÄT DORTMUND REIHE COMPUTATIONAL INTELLIGENCE COLLABORATIVE RESEARCH CENTER 531

Runtime Analysis of the (1 + (λ, λ)) Genetic Algorithm on Random Satisfiable 3-CNF Formulas

UNIVERSITY OF DORTMUND

First hitting time analysis of continuous evolutionary algorithms based on average gain

On the convergence rates of genetic algorithms

Evolutionary Computation

Lecture 9 Evolutionary Computation: Genetic algorithms

Rigorous Analyses for the Combination of Ant Colony Optimization and Local Search

Interplanetary Trajectory Optimization using a Genetic Algorithm

Gene Pool Recombination in Genetic Algorithms

Efficient Non-domination Level Update Method for Steady-State Evolutionary Multi-objective. optimization

UNIVERSITY OF DORTMUND

Expected Running Time Analysis of a Multiobjective Evolutionary Algorithm on Pseudo-boolean Functions

Runtime Analysis of Evolutionary Algorithms: Basic Introduction 1

Optimizing Linear Functions with Randomized Search Heuristics - The Robustness of Mutation

Adaptive Generalized Crowding for Genetic Algorithms

k-protected VERTICES IN BINARY SEARCH TREES

The Nottingham eprints service makes this work by researchers of the University of Nottingham available open access under the following conditions.

Runtime Analysis of Mutation-Based Geometric Semantic Genetic Programming on Boolean Functions

1. Computação Evolutiva

Improved time complexity analysis of the Simple Genetic Algorithm

REIHE COMPUTATIONAL INTELLIGENCE S O N D E R F O R S C H U N G S B E R E I C H 5 3 1

The (1+) evolutionary algorithm with self-adjusting mutation rate

Fitness Inheritance in Multi-Objective Optimization

Lecture 4 October 18th

Estimation-of-Distribution Algorithms. Discrete Domain.

Defining Locality as a Problem Difficulty Measure in Genetic Programming

Supplementary lecture notes on linear programming. We will present an algorithm to solve linear programs of the form. maximize.

UNIVERSITY OF DORTMUND

CSC 4510 Machine Learning

Convex Functions and Optimization

Compute the Fourier transform on the first register to get x {0,1} n x 0.

METHODS FOR THE ANALYSIS OF EVOLUTIONARY ALGORITHMS ON PSEUDO-BOOLEAN FUNCTIONS

Quad-trees: A Data Structure for Storing Pareto-sets in Multi-objective Evolutionary Algorithms with Elitism

UNIVERSITY OF DORTMUND

Runtime Analyses for Using Fairness in Evolutionary Multi-Objective Optimization

Analysis of Evolutionary Algorithms for the Longest Common Subsequence Problem

Evolutionary Programming Using a Mixed Strategy Adapting to Local Fitness Landscape

On the Approximation Ability of Evolutionary Optimization with Application to Minimum Set Cover: Extended Abstract

2-bit Flip Mutation Elementary Fitness Landscapes

Stochastic Process (ENPC) Monday, 22nd of January 2018 (2h30)

On the Topology of Genetic Algorithms

Running Time Analysis of Multi-objective Evolutionary Algorithms on a Simple Discrete Optimization Problem

Convergence of Ant Colony Optimization on First-Order Deceptive Systems

Course 212: Academic Year Section 1: Metric Spaces

September Math Course: First Order Derivative

IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. XX, NO.X, XXXX 1

Pseudorandom Generators

Introduction to Machine Learning (67577) Lecture 3

Exploration of population fixed-points versus mutation rates for functions of unitation

Algorithms: COMP3121/3821/9101/9801

NP Completeness and Approximation Algorithms

UNIVERSITY OF DORTMUND

Experimental Supplements to the Theoretical Analysis of EAs on Problems from Combinatorial Optimization

8. Prime Factorization and Primary Decompositions

arxiv: v1 [cs.dm] 29 Aug 2013

Running time analysis of a multi-objective evolutionary algorithm on a simple discrete optimization problem

Introduction to Walsh Analysis

Parallel Genetic Algorithms

Local Search (Greedy Descent): Maintain an assignment of a value to each variable. Repeat:

Lecture 9 Classification of States

Where do pseudo-random generators come from?

Transcription:

Average Drift Analysis and Population Scalability Jun He and Xin Yao Abstract This paper aims to study how the population size affects the computation time of evolutionary algorithms in a rigorous way. The computation time of evolutionary algorithms can be measured by either the number of generations (hitting time) or the number of fitness evaluations (running time) to find an optimal solution. Population scalability is the ratio of the expected hitting time between a benchmark algorithm and an algorithm using a larger population size. Average drift analysis is introduced to compare the expected hitting time of two algorithms and to estimate lower and upper bounds on the population scalability. Several intuitive beliefs are rigorously analysed. It is proven that () using a population sometimes increases rather than decreases the expected hitting time; (2) using a population cannot shorten the expected running time of any elitist evolutionary algorithm on any unimodal function on the time-fitness landscape, however this statement is not true in terms of the distance-based fitness landscape; (3) using a population cannot always reduce the expected running time on deceptive functions, which depends on whether the benchmark algorithm uses elitist selection or random selection. Index Terms evolutionary algorithm, computation time, population size, fitness landscape, drift analysis. I. INTRODUCTION Population is one of the most important features of evolutionary algorithms (EAs). A wide range of approaches is available to design population-based EAs. Using a population delivers many benefits []. The study of the relationship between the performance of an EA and its population size can be traced back to early 990s. For example, Goldberg et al. [2] presented a population sizing equation to show how a large population size helps an EA to distinguish between good and bad building blocks on some test problems. Mühlenbein and Schlierkamp-Voosen [3] studied the critical (minimal) population size that can guarantee the convergence to the optimum. Arabas et al. [] proposed an adaptive scheme for controlling the population size, and the effectiveness of the proposed scheme was validated by an empirical study. Eiben et al. [5] reviewed various techniques of parameter controlling for EAs, where the adjustment of population size was considered as an important research issue. Harik et al. [6] linked the population size to the quality of solution by the analogy between one-dimensional random walks and EAs. The theoretical analysis of the impact of the population size on the computation time of EAs starts in early 2000s [7]. There has been an increasing interest in rigorously analysing Manuscript received xx xx xxxx This work was supported by the EPSRC under Grant Nos. EP/I00909/, EP/I00297/ and EP/K00523/, and by NSFC under Grant No. 6329302. XY was also supported by a Royal Society Wolfson Research Merit Award. Jun He is with Department of Computer Science, Aberystwyth University, Aberystwyth, SY23 3DB, UK. Email: jun.he@aber.ac.uk Xin Yao is with CERCIA, School of Computer Science, University of Birmingham, Birmingham B5 2TT, UK. Email: x.yao@cs.bham.ac.uk the relationship between the computation time of an EA and its population size. The computation time of an EA can be measured by either the expected hitting time or the expected running time. The theoretical studies on this topic can be classified into two directions. One direction aims to estimate a bound on the computation time of EAs as a function of the population size. This direction belongs to the time complexity analysis of EAs. Drift analysis and tail inequalities are often used for estimating the time bound. This direction may be called a bound-based study. A lot of work has done along this direction. The earliest one was conducted by Jansen et al. [] who first obtained the cut-off point for a ( + λ) EA on three pseudo-boolean functions, Leading-Ones, One-Max and Suf-Samp. Jägersküpper and Witt [9] analysed how the running time of a (µ+) EA on the Sphere function scales up with respect to µ. Witt [0] proved theoretically that the running time of a (µ+) EA on a specific pseudo-boolean function is polynomial with an overwhelming probability, when µ is large enough. Storch [] presented a rigorous runtime analysis of the choice of the population size with respect to a (µ + ) EA on several pseudo-boolean functions. Yu and Zhou [2] investigated the expected hitting time of (λ + λ) EAs when λ = and λ = n (where n is the problem input size) on the trap problem. Oliveto et al. [3] presented a runtime analysis of both ( + λ) and (µ+) EAs on some instances of the vertex covering problem. Friedrich et al. [] analysed the running time of a (µ + ) EA with diversity-preserving mechanisms on the Two-Max problem. Chen et al. [5] obtained an upper bound on the hitting time of (λ + λ) EAs on Leading-Ones and One-Max problems. Lässig and Sudholt [6] presented a running time analysis of a ( + λ) EA with an adaptive offspring size λ on several pseudo-boolean functions. Rowe and Sudholt [7] discussed the running time of ( + λ) EAs in terms of the offspring population size on unimodal functions. Doerr and Künnemann [] analysed the time bound of ( + λ) EAs for optimizing linear pseudo-boolean functions. Doerr and Künnemann [9] showed that (+λ) EAs with even very large offspring populations does not reduce the runtime significantly on the Royal Road function. Oliveto and Witt [20] presented a rigorous running time analysis of the well-known Simple Genetic Algorithm for One-Max. Gießen and Witt [2] studied the relation between the population size and mutation strength for a ( + λ) EA on One-Max. Another direction aims to calculate a ratio, named population scalability, which is described as follows: expected hitting time of a benchmark EA expected hitting time of an EA using a population. () This direction may be called a ratio-based study. As one of the earliest analyses, He and Yao [7] investigated how the

2 population scalability of EAs varies as the population size changes on two simple functions (One-Max and Deceptive). In that paper, EAs are assumed to be run on a hypothetical parallel computer, that is, to assign each individual on one processor. If the communication cost is ignored, the population scalability is equivalent to the speedup in parallel computation. The link between population scalability and parallelism was further discussed in [22]. However, since calculating the population scalability is not an easy job, no further development has been made since then. This paper belongs to the ratio-based study. It is significantly different from the bound-based study. The bound-based study focuses on an asymptotic time bound as a function of the population size. It does not calculate the population scalability and will not answer whether a (2+2) EA is faster than a (+) EA. The ratio-based study aims to calculate the population scalability and will answer whether a (2 + 2) EA is faster than a ( + ) EA. But it is not necessary to estimate an asymptotic time bound. Compared with previous work on the analysis of populationbased EAs, the current paper has two novelties. ) Average drift analysis is presented as a tool of comparing the expected hitting time of two EAs and studying population scalability. The approach used in [7] is based on the fundamental matrix of absorbing Markov chains. It is hard to calculate the expected hitting time through the fundamental matrix. But using average drift analysis, it is possible to estimate population scalability without calculating the expected hitting time. This is an important improvement in the analysis tool. 2) The scalability threshold replaces the cut-off point. The population threshold is the minimal population size at which the running time of an EA using a larger population size is greater than that of the benchmark EA. The cut-off point [] is the maximize population size at which the running time of the EA is in the same order as that of the benchmark EA. Let s show the advantage of population scalability by an example: (+λ) EAs (using bitwise mutation and elitist selection) for solving the One-Max ( problem. ) According to [], the cut-off point is Θ (ln n)(ln ln n) ln ln ln n. This means when the population size λ is smaller than the cut-off point, the running time of the (+λ) EA is in the same order as that of the (+) EA but different by a constant factor. The constant could be =, > or <. Therefore the cut-off point does not answer the question whether the expected ( running ) time of a ( + λ) EA (where 2 λ = O (ln n)(ln ln n) ln ln ln n ) is smaller or larger than that of the ( + ) EA. However, according to Proposition and its discussion in this paper, the scalability threshold is 2. This means that the running time of the (+λ) EA (for any λ 2) is larger than that of the ( + ) EA. Therefore the scalability threshold is more accurate than the cut-off point. With the help of average drift analysis, this paper analyses the following intuitive beliefs in a rigorously way. ) Using a population always reduces the expected hitting time (not running time) of an EA to find an optimal point. 2) Using a population cannot shorten the expected running time of an elitist EA on unimodal functions. 3) Using a population can reduce the expected running time of an EA on deceptive functions. The paper is organised as follows: Section II defines population scalability. Section III presents drift analysis for population scalability. Section IV analyses scenario : using a population does not reduce the hitting time. Section V analyses scenario 2: using a population reduces the hitting time, but not the running time. Section VI investigates scenario 3: using a population reduces the running time. Section VIII concludes the paper. A. Evolutionary algorithms II. POPULATION SCALABILITY Consider the problem of maximizing a function f(x) where x S and S is a finite set. A point in S is called a solution or an individual. A population consists of one or more individuals. A (µ + λ) EA is described in Algorithm. The stopping criterion is that the EA halts once an optimal solution is found. The criterion is used for the sake of analysis because our interest is the first hitting time (when the EA finds an optimal solution for the first time). If Φ t includes an optimal solution, assign Φ t = Φ t+ = Φ t+2 = for ever. Algorithm A (µ + λ) EA where µ, λ : initialise a population Φ 0 consisting of µ individuals (solutions) and t 0; 2: evaluate the fitness of individuals in Φ 0 ; 3: while Φ t does not include an optimal solution do : mutate (or crossover) individuals in Φ t and generate a children population Ψ t consisting of λ individuals; 5: evaluate the fitness of individuals in Ψ t ; 6: probabilistically select µ individuals from Φ t Ψ t as Φ t+ ; 7: t t + ; : end while The sequence {Φ t ; t = 0,, } can be modelled by a Markov chain [23]. Each generation of the EA consists of two steps: to generate new individuals by mutation or crossover and to select individuals for next generation, mutation (or crossover) Φ t Ψ t Φ t selection Φ t+. Let P denote the set of all populations, P opt the set of populations including an optimal solution and P non the set of populations without an optimal solution. The transition from Φ t to Ψ t can be represented using mutation (or crossover) probabilities: P m (X, Y ) def = P (Ψ t = Y Φ t = X), X, Y P, (2) where Φ t, Ψ t are random variables representing the tth generation population and its children population. X, Y are their values taken from P.

3 The transition from Φ t and Ψ t to Φ t+ can be represented using selection probabilities: P s (X, Y, Z) def =P (Φ t+ = Z Φ t = X, Ψ t = Y ). (3) The transition from Φ t from Φ t+ can be represented using transition probabilities: P (X, Y ) def = Pr(Φ t+ = Y Φ t = X). () The hitting time is the number of generations of an EA to find an optimal solution for the first time. Definition : Given Φ 0 = X, the expected hitting time of an EA is defined by + g(x) def = Pr(Φ t P non ). (5) t=0 If the initial population Φ 0 is chosen according to a probability distribution over P, the expected hitting time is given by g(φ 0 ) def = X P g(x) Pr(Φ 0 = X). The expected running time of a (µ + λ) EA is the expected number of fitness evaluations, which equals to µ + λg(φ 0 ). For the sake of simplicity, we always omit the first term µ, which is the number of fitness evaluations in initialization. If genetic operators don t change in time, the sequence {Φ t ; t = 0,, } can be modelled by a homogeneous Markov chain. According to the fundamental matrix theorem [2, Theorem.5], the expected hitting time of an EA can be calculated from transition probabilities. Theorem : If the population sequence {Φ t, t = 0,, } is a homogeneous Markov chain and converges to P opt, that is, lim t + Pr(Φ t P opt ) =, then the expected hitting time g(x) satisfies a linear equation system: { g(x) = 0, if X Popt, Y P P (X, Y ) (g(x) g(y )) =, if X / Pnon. (6) The fundamental matrix theorem is useful in analysing elitist EAs [7], [23], [25]. However, its disadvantage is the difficulty of solving the above linear equation system. B. Population scalability The population scalability is defined as the ratio of the expected hitting time between a benchmark EA and an EA with a larger population size. In this paper, the benchmark is a (+) EA using mutation and selection operators. Other types of EAs may play the role of a benchmark too. For example, a (2 + 2) EA could be chosen as a benchmark when studying EAs with crossover. But we will not discuss them here. Definition 2: Given a ( + ) EA and a (µ + λ) EA that exploit an identical mutation operator to optimise the same fitness function, let Φ (+) 0 and Φ (µ+λ) 0 denote their corresponding initial populations, then the population scalability is defined by P S(µ + λ Φ (+) 0, Φ (µ+λ) 0 ) def = g(+) (Φ (+) 0 ) g (µ+λ) (Φ (µ+λ) 0 ), (7) where the superscripts (+) and (µ+λ) are used to distinguish the ( + ) EA and (µ + λ) EA. An essential part of the definition above is that both EAs must adopt identical mutation operators. This ensures that the comparison is meaningful. Nonetheless, it is impossible for the selection operators to be identical. Indeed even if the selection operators are of the same type, for example roulette wheel selection, the conditional probabilities determining the actual selection operators are never identical under distinct population sizes. Obviously the value of population scalability relies on initial populations. Due to the use of a population, Φ (µ+λ) 0 may contain several individuals some of which are different from Φ (+) 0. For the sake of comparison, we restrict our discussion to identical initialization, that is, for the (+) EA, Φ (+) 0 = x and for the (µ + λ) EA, Φ (µ+λ) 0 = (x,, x). In this case, P S(µ + λ Φ (+) 0, Φ (µ+λ) 0 ) is denoted by P S(µ+λ x) in short. There exist other types of initialization but we will not discussed them here. The notion of population scalability is similar to that of the speedup widely used in analyzing parallel algorithms. The speedup of parallel EAs have been studied through experiments [26], [27], [2]. If each individual is assigned to a processor, then EAs turn into parallel EAs. Under this circumstance, population scalability is equivalent to speedup if ignoring the communication cost. Hence population scalability is called speedup on a hypothetical parallel computer in [7]. The following questions are essential when studying population scalability. ) Given a λ 2 or µ 2, is the population scalability P S(µ + λ x) >? If it is, we may assign each individual to a processor in a parallel computing system and then the CPU computation time of the (µ + λ) EA is less than that of the ( + ) EA (if ignoring the communication cost). 2) Given a λ 2 or µ 2, is the population scalability P S(µ + λ x) > λ? If it is, then the CPU computation time of the (µ + λ) EA on a computer is less than that of the ( + ) EA. 3) Where are the smallest population sizes (µ, λ) such that the expected running time of the (µ + λ) EA is larger than that of the ( + ) EA? We call this point the scalability threshold, which satisfies { min{µ : P S(µ + λ x) > λ}, () min{λ : P S(µ + λ x) > λ}. In general, the scalability threshold is not a single point but a Pareto front due to minimizing both population sizes µ and λ simultaneously. However, in a ( + λ) or (λ + λ) EA, the scalability threshold is a single point. III. AVERAGE DRIFT ANALYSIS AND TIME-FITNESS A. Average drift analysis LANDSCAPE Average drift analysis is a variant of drift analysis for estimating the expected hitting time of EAs. The idea of

average drift was first used by Jägersküpper who considered the average drift of a ( + ) EA on linear functions and provided a delicate analysis of the running time of the ( + ) EA [29]. Nevertheless his work was restricted to the ( + ) EA and linear functions. Even the term of average drift did not appear in [29] but was first adopted by Doerr [30] for introducing Jägersküpper s work [29]. In this section, the average drift is formally defined and then general average drift theorems are presented. In drift analysis, a distance function d(x) is used to measure how far a population X is away from the optimal set P opt. It is a non-negative function such that 0 < d(x) < + for any X P non and d(x) = 0 for X P opt. Drift is used to measure the progress rate of a population moving towards the optimal set per generation. Definition 3: Given a population X, the pointwise drift at X is (X) def = (d(x) d(y )) Pr(Φ t+ = Y Φ t = X). (9) Y P Given a generation t, the average drift at t is { def 0, if Pr(Φt P non)=0, t = X Pnon (X) Pr(Φ t =X) Pr(Φ t Pnon), otherwise. (0) The following theorem provides an approach to estimating a lower bound on the expected hitting time. It is a variation of [3, Theorem ]. Theorem 2: Provided that the population sequence {Φ t, t = 0,, } converges to P opt where Φ 0 satisfies Pr(Φ 0 P non ) > 0. Given a distance function d(x), if for any t and any Φ t such that Pr(Φ t P non ) > 0, the average drift t c where c > 0, then the expected hitting time g(φ 0 ) d(φ 0 )/c, where d(φ 0 ) def = d(x) Pr(Φ 0 = X). X P Furthermore if for at least one t, the average drift t < c, then g(φ 0 ) > d(φ 0 )/c. Proof: Without loss of generality, let c =. From the condition t for any t and any Φ t such that Pr(Φ t P non ) > 0, we have Pr(Φ t P non ) (X) Pr(Φ t = X) X P non X Pnon d(x) Pr(Φt=X) Y d(y ) Pr(Φt+=Y ). () Pnon Summing the term Pr(Φ t P non ) from t = 0 to k, we get k Pr(Φ t P non ) t=0 k t=0 ( X Pnon d(x) Pr(Φt=X) Y d(y ) Pr(Φt+=Y )) Pnon = X Pnon d(x) Pr(Φ0=X) Y Pnon d(y ) Pr(Φ k+=y ). (2) Notice that Y P non d(y ) Pr(Φ k+ = Y ) max X P d(x) Y P non Pr(Φ k+ = Y ) = max X P d(x) Pr(Φ k+ P non ). (3) Since the EA is convergent: lim k + Pr(Φ k+ P non ) = 0, then from Inequality (3) we have lim d(y ) Pr(Φ k+ = Y ) = 0. () k + Y P non Applying the above result to Inequality (2) (let k + ), we get + g(φ 0 ) = Pr(Φ t P non ) t=0 X P non d(x) Pr(Φ 0 = X) = d(φ 0 ), (5) which gives the desired result. If for some t, the average drift t <, then Inequality (2) is strict for any k t and Inequality (5) is strict too. Similarly, the theorem below provides an approach to estimating an upper bound on the expected hitting time. It is a variation of [3, Theorem ]. Its proof is similar to that of the above theorem. Theorem 3: Provided that population sequence {Φ t, t = 0,, } converges to P opt where Φ 0 satisfies Pr(Φ 0 P non ) > 0. Given a distance function d(x), if for any t and any Φ t such that Pr(Φ t P non ) > 0, the average drift t c where c > 0, then the expected hitting time g(φ 0 ) d(φ 0 )/c. Furthermore if for at least one t, the average drift t > c, then g(φ 0 ) < d(φ 0 )/c. Pointwise drift theorems are corollaries of average drift theorems, because it requires a stronger condition on the pointwise drift: (X) c (or c) for any X P non. It implies the average drift t c (or c) for any Φ 0 P non. Theorem : [23, Theorem 2] Provided that population sequence {Φ t, t = 0,, } converges to P opt. Given a distance function d(x), if for any X P non, the pointwise drift (X) c (where c > 0), then for any initial population X 0 P non, the expected hitting time g(x 0 ) d(x 0 )/c. Theorem 5: [23, Theorems 3] Provided that population sequence {Φ t, t = 0,, } converges to P opt. Given a distance function d(x), if for any X P non, the pointwise drift (X) c where c > 0, then for any initial population X 0 P non, the expected hitting time g(x 0 ) d(x 0 )/c. The average drift analysis provides a useful tool for comparing the expected hitting time of two EAs. Its idea is simple. One EA is taken as the benchmark and its expected hitting time is used to define a distance function for the other EA. Then the average drift of the other EA is estimated and then its expected hitting time is bounded using average drift analysis. Theorem 6: Given two EAs A and B to optimise the same fitness function, let {X t, t = 0, } and {Y t, t = 0,, } denote their population sequences respectively. For algorithm B, define a distance function d B (X) such that d B (Y 0 ) = g A (X 0 ), where g A (X 0 ) is the expected hitting time of algorithm A starting at X 0. If for any t and any Y t such that Pr(Y t P non ) > 0, average drift B t c where c > 0, then the expected hitting time of algorithm B satisfies that g B (Y 0 ) g A (X 0 )/c. Furthermore if for at least one t 0, B t < c, then g B (Y 0 ) > g A (X 0 )/c. Proof: It is a direct corollary of Theorem 2.

5 Theorem 7: Given two EAs A and B to optimise the same fitness function, let {X t, t = 0, } and {Y t, t = 0,, } be their population sequences respectively. For algorithm B, define a distance function d B (X) such that d B (Y 0 ) = g A (X 0 ), where g A (X 0 ) is the expected hitting time of algorithm A starting at X 0. If for any t and any Y t such that Pr(Y t P non ) > 0, average drift B t c where c > 0, then g B (Y 0 ) g A (X 0 )/c. Furthermore if for some t 0, B t > c, then g B (Y 0 ) < g A (X 0 )/c. Proof: It is a direct corollary of Theorem 3. B. Average drift analysis for population scalability Average drift analysis for estimating the population scalability is based on an simple idea. Given a benchmark EA and another EA using a larger population size, we assume that both EAs start at the same point with an equal distance to the optimal set. If at each generation, the average drift of the other EA is 0 times that of the benchmark EA, then the expected hitting time of the other EA will be /0 of that of the benchmark EA. Thus the population scalability is 0. This simple idea can be formalised as follows. Consider a (+) EA and a (µ+λ) EA (where λ 2) that exploit an identical mutation operator to optimise the same fitness function. Provided that Φ (+) 0 = x 0 and Φ (µ+λ) 0 = (x 0,, x 0 ) for some x 0 P non, for the (µ + λ) EA, define a distance function d(x) such that d (µ+λ) (x 0,, x 0 ) = g (+) (x 0 ). The first theorem establishes a sufficient condition for estimating the upper bound on population scalability. Thanks to average drift analysis, there is no requirement that the population sequence is a Markov chain. Theorem : If for all t 0 and any Φ (µ+λ) t such that Pr(Φ (µ+λ) t P non ) > 0, average drift (µ+λ) t c where c > 0, then P S(µ + λ x 0 ) c. Furthermore if for at least (µ+λ) one t 0, average drift t < c, then P S(µ+λ x 0 ) < c. Proof: According to Theorem 6, the expected hitting time satisfies: g (µ+λ) (x 0 ) g (+) (x 0 )/c. Then we have P S(µ + (µ+λ) λ x 0 ) c. If for at least one t 0, average drift t < c, then according to Theorem 6, g (µ+λ) (x 0 ) g (+) (x 0 )/c and then P S(µ + λ x 0 ) < c. Similarly, the second theorem establishes a sufficient condition for estimating the lower bound on population scalability. Its proof is the almost the same as the that of the above theorem except using Theorem 7. Theorem 9: If for all t 0 and any Φ (µ+λ) t such that Pr(Φ (µ+λ) t P non ) > 0, average drift (µ+λ) t c where c > 0, then P S(µ + λ x 0 ) c. Furthermore if for at least (µ+λ) one t 0, average drift t > c, then P S(µ+λ x 0 ) > c. C. Time-fitness landscape In this paper, unimodal and multi-modal functions are established upon the time-fitness landscape, a concept introduced in [32]. It aims at describing the fitness landscape related to a general search space, which is a finite set. Definition : Given a ( + ) elitist EA for maximizing a function f(x), its time-fitness landscape is the set of pairs (g(x), f(x)), where g(x) is the expected hitting time of the ( + ) EA starting at x. The neighbour of x includes two points: the point y such that g(y ) is the closest to g(x) from the direction g(y) < g(x), and the point y 2 such that g(y 2 ) is the closest to g(x) from the direction g(y) > g(x). The time-fitness landscape is completely different from traditional ones based on a distance. The former is related to a ( + ) EA, but the latter usually not. Let s show the difference by unimodal functions. A function is called unimodal if every non-optimal point has a neighbour with a strictly better fitness [7]. Traditionally the definition of the neighbour relies on a distance. For example, if the search space is the set of binary strings, the neighbour of a point x includes all points y with Hamming distance from x [7]. But such a definition is applicable to a finite set because the distance is unknown or not defined. Therefore for a general finite set, unimodal functions are defined on the time-fitness landscape instead [32]. Definition 5: Let S = {s 0, s,, s K } and f a fitness function such that f(s 0 ) > f(s ) > > f(s K ). Given a ( + ) elitist EA to maximise a function f, f is called unimodal to the (+) EA if g (+) (s ) < < g (+) (s K ). A unimodal time-fitness landscape is visualised in Fig.. f(x) s 0 s g (+) (x) s K Fig.. A unimodal time-fitness landscape. The x axis is the expected hitting time of the ( + ) EA. The y axis is the fitness function. According to [32], unimodal functions to a ( + ) elitist EA are the easiest among all fitness functions with the same optimum at s 0. Furthermore, according to [32], for any fitness function f(x), we can construct a ( + ) elitist EA to which f is the easiest (unimodal). Therefore any fitness function can be unimodal to a ( + ) elitist EA. Here are two instances of unimodal functions. The TwoMax function [32] below is unimodal under Hamming distance. s K f(x) = max{ x, n x }, x {0, } n, (6) where x denotes the number of -valued bits. It is unimodal to the following ( + ) elitist EA with p = n [32]. Bitwise Mutation p. Flip each bit independently with flipping probability p. Elitist Selection. Select the best individual from Φ t Ψ t. The Two Needles in the Haystack function [32] below is also unimodal to the above ( + ) elitist EA with p = 2, although it is regarded as a plateau function under Hamming distance. {, if x = 0 or x = n; f(x) = (7) 0, otherwise.

6 IV. SCENARIO : USING A POPULATION CANNOT REDUCE HITTING TIME A. Case study : two-paths-i functions It is an intuitive belief that using a population will reduce the number of generations to find an optimal solution. The following case study shows this belief is not always true. Before the case study, the concept of path [33] is revisited. Given a ( + ) EA for maximizing f(x), a path is a sequence of points {x 0 x x k } such that P (x i, x i ) > 0 for i =,, k. The path is denoted by Path(x 0, x,, x k ). The case study is about two-paths-i functions which are defined as below. Definition 6: Let S = {s 0, s,, s K+L, s K+L+ } where L > K and f a fitness function such that f(s 0 ) > f(s K+ ) > f(s K+2 ) > > f(s K+L ) >f(s ) > f(s 2 ) > > f(s K ) > f(s K+L+ ). () Given a ( + ) elitist EA to maximise a function f, f is called a two-paths-i function to the ( + ) EA if there exist two paths to the optimum: Path (s K+L+, s K,, s, s 0 ) and Path 2 (s K+L+, s K+L,, s K+, s 0 ) such that for k =,, K and k = K + 2,, K + L, mutation probabilities P m (s k, s k ) = ; for k = K +, mutation probability P m (s k, s 0 ) = ; for k = K +L+, mutation probabilities P (s k, s K ) = p and P m (s k, s K+L ) = p where 0 < p < ; for any other i, j, P m (s i, s j ) = 0. Fig. 2 visualises a two-paths-i time-fitness landscape. f(x) s 0 s K+ s s K sk+l+ s g (+) (x) s K+L Fig. 2. A two-paths-i time-fitness landscape. The x axis represents the expected hitting time of the ( + ) EA. The y axis is the fitness function. Consider a ( + λ) EA (where λ 2) for maximizing a two-paths-i function function. Mutation. Mutation probabilities are identical to those in the ( + ) EA. Generate λ children. Elitist Selection. Select the best individual from Φ t Ψ t. The theorem below shows that using a population will increase the expected hitting time if the EA starts at s K+L+. Proposition : Given the ( + ) elitist EA and a ( + λ) EA (where λ 2) for maximizing a two-paths-i function, let Φ (+) 0 = Φ (+λ) 0 = s K+L+, then P S( + λ s K+L+ ) <. The scalability threshold is 2. Proof: For the ( + λ) EA, let its distance function d (+λ) (x) = g (+) (x). Consider the pointwise drift at s K+L+. There are two potential events at S K++. ) The ( + λ) EA moves from s K+L+ to s K. This event happens if and only if all children are s K. The probability for the event happening is p λ. 2) The ( + λ) EA moves from s K+L+ to s K+L. This event happens if and only if at least one child is s K+L. The probability for the event happening is p λ. We calculate the pointwise drift at s K+L+ as follows: (+λ) (s K+L+ ) =( p λ )(g (+) (s K+L+ ) g (+) (s K+L )) + p λ (g (+) (s K+L+ ) g (+) (s K )) =( p λ )[ + pk + ( p)l L] + p λ [ + pk + ( p)l K] = + (p p λ )(K L) <. (9) (since L > K, p (0, ) and λ 2) For any other non-optimal s {s,, s K+L }, we calculate the pointwise drift (+λ) (S k ) = from P (s k, s k ) = (except k = K + ) and P (s K+,0 ) =. Since Φ (+λ) (+λ) 0 = s K+L+, average drift 0 <. For any t, Φ (+λ) t has left the point s K+L+, then average drift (+λ) t =. According to Theorem, we get that P S(+λ s K+L+ ) <. It is easy to understand the proposition. There are two paths towards the optimum: short and long. If using a population, the long path is more likely to be chosen than the short path. Then the expected hitting time is increased. Example : Consider an instance of two-paths-i functions. Let x {0, } n and x denote its number of -valued bits. n, if x = 0 or x = n, f(x) = x, if x θ where θ = 2, x, otherwise. (20) There are two optima: x = 0, n. Let x 0 be a string such that x 0 = 2. It takes the minimum value of the fitness. A ( + λ) EA uses adaptive mutation and elitist selection for solving the above problem. Adaptive mutation. ) When t = 0, flip one of -valued bits with the probability 0.5, otherwise flip one of 0-valued bits. In this way, generate λ children as Ψ 0. 2) When t, if f(φ t ) > f(φ t ) and Φ t is generated by flipping a 0-valued bit (or -valued) in Φ t, flip a 0-valued bit (or -value bit). Then generate λ children as Ψ t. 3) When t, if f(φ t ) = f(φ t ), either flip one of -valued bits with probability 0.5 or flip one of 0-valued bits. Then generate λ children as Ψ t. Elitist Selection. Select the best individual from Φ t Ψ t as Φ t+. f is a unimodal function under Hamming distance but a twopaths-i function (multi-modal) to the ( + ) EA on the timefitness landscape. When the (+) EA starts at x = 2, there are two paths towards the optima x = 0, n. The short path is x = 2 0. The long path is x = 2 3 n. Table I shows that using a population increases the expected hitting time.

7 TABLE I EXPERIMENTAL RESULTS FOR EXAMPLE AVERAGED OVER 000 RUNS. n = 000. THE EA STARTS AT x 0 WITH x 0 = 2. population size 3 5 7 9 hitting time 9 72 96 990 997 running time 9 266 05 6930 973 B. Case study 2: unimodal functions This subsection presents another case study to show that using a population can not reduce the expected hitting time. The case study discusses a (µ + ) elitist EA (where µ ) with global mutation for maximizing any unimodal function. Global Mutation. Choose one individual from Φ t at random and generate a child by mutation. Mutation probability P m (s i, s j ) > 0 for any i and j. Elitist Selection. If the child s fitness is better than that of one or more parents, then the child will replace one of these parents at random. The proposition below asserts that using a population increases the expected hitting time of the (µ + ) EA. Proposition 2: Given the ( + ) EA and a (µ + ) EA (where µ 2) for maximizing a unimodal function, for any x {s 2,, s K }, let Φ (+) 0 = x and Φ (µ+) 0 = (x,, x), then P S(µ + x) <. The scalability threshold is 2. Proof: For the ( + ) EA, choose g (+) (x) to be its distance function. According to Theorem, the pointwise drift satisfies (+) (s k ) = for any s k {s,, s K }. Since selection is elitist and the function is unimodal, we have (+) (s k ) = k Pm(s k,s l )(g (+) (s k ) g (+) (s l ))=. (2) For the (µ+) EA, define its distance function d (µ+) (X) = min{g (+) (x) : x X}. Let X be a parent population: X = (s (X),, s µ(x) ) P non in which the best parent is s k. Let their child be s l. Since the (µ + ) EA adopts elitist selection and the fitness function is unimodal, the pointwise drift satisfies (µ+) (X) = k Pm(X,s l)(g (+) (s k ) g (+) (s l )). (22) The probability of mutating a parent s i(x) to the child s l (where l < k) is P m (s i(x), s l ). Since each parent is chosen for mutation at random, we have Then we get P m (X, s l ) P m (X, s l ) = µ µ P m (s i(x), s l ). i= ( ) g (+) (s k ) g (+) (s l ) = µ µ i= P m(s i(x), s l )(g (+) (s k ) g (+) (s l )). Since l < k i(x), according to the definition of unimodal functions, we have g (+) (s l ) < g (+) (s k ) g (+) (s i(x) ), and then P m (X, s l ) ( ) g (+) (s k ) g (+) (s l ) µ µ i= P m(s i(x), s l )(g (+) (s i(x) ) g (+) (s l )) The pointwise drift satisfies (µ+) (X) = µ k µ i= Pm(s i(x),s l )(g (+) (s k ) g (+) (s l )) µ µ i= k Pm(s i(x),s l )(g (+) (s i(x) ) g (+) (s l )). (23) ) Case : for all parents in X, s i(x) = s k. According to Equality (2), we have k P m(s k, s l )(g (+) (s k ) g (+) (s l )) =. Then the pointwise drift satisfies (µ+) (X) = µ =. (2) µ i= 2) Case 2: for at least one parent in X, s i(x) s k. Since s k is the best parent in X, the indexes satisfy i(x) > k. Thanks to global mutation, we have P m (s i(x), s l ) > 0 for any i(x), and then k Pm(s i(x),s l )(g (+) (s i(x) ) g (+) (s l )) < i(x) P m(s i(x),s l )(g (+) (s i(x) ) g (+) (s l )) = (+) (s i(x) ) =. (25) The last equality = comes from Equality (2). Then the pointwise drift satisfies (µ+) (X) < µ =. (26) µ Since Φ 0 = (s k,, s k ) where k 2, the average drift (µ+) 0 =. When t =, the probability of Φ P non including two different non-optimal points is always greater (µ+) than 0 due to global mutation. Thus the average drift <. When t 2, the average drift (µ+) t. According to Theorem, we get P S(µ + s k ) <. Here is an explanation of this proposition. For unimodal functions, the higher the fitness of a point is, the closer to the optimal set the point is (Fig. ). Given a parent population in the (µ + ) EA, a good strategy is to mutate the best parent in the population. Unfortunately, a population may include some parent which is worse than the best. If a worse parent is chosen to mutate, it will increase he expected hitting time. Example 2: Consider the TwoMax function [32]. Let x {0, } n and x denote the number of -valued bits. i= f(x) = max{ x, n x }. (27) There are two optima: x = 0, n. A (µ + ) elitist EA is used for solving the maximization problem. Bitwise Mutation. Choose one individual from µ parents at random. Flip each bit with a probability /n. Elitist Selection. If the child s fitness is better than that of one or more parents, then the child will replace one of these parents at random. f is unimodal to the ( + ) EA [32]. Table II shows that using a population increases the expected hitting time.

TABLE II EXPERIMENTAL RESULTS FOR EXAMPLE 2 AVERAGED OVER 000 RUNS. n = 200. THE EA STARTS AT x 0 WITH x 0 = 00. population size 3 5 7 9 hitting time 2523 2650 30 377 355 running time 2523 2650 30 377 355 V. SCENARIO 2: USING A POPULATION CAN REDUCE HITTING TIME BUT NOT RUNNING TIME A. Case study 3: unimodal functions Let s reinvestigate the intuitive belief that using a population can reduce the expected hitting time of elitist EAs for maximizing unimodal functions. Although this belief is not true for the (µ + ) EA, it is still true for the (λ + λ) EA with global mutation and elitist selection. Consider a (λ + λ) EA (λ ) using elitist selection and global mutation for maximising a unimodal function. Global Mutation. Mutation probability P m (s i, s j ) > 0 for any i, j. Each individual in Φ t generates a child. Elitist Selection. Probabilistically select λ individuals from Φ t Ψ t, while the best individual is always selected. First we prove an inequality which will be used later. Lemma : Given a i > 0, b i > 0, c i > 0 where i = 0,,, k such that j i=0 a i > j i=0 b i, j = 0,, k and c 0 > c > > c k, it holds k i=0 a ic i > k i=0 b ic i. Proof: From the conditions a 0 + + a j > b 0 + + b j and c 0 > c > > c k, we have k i=0 (a i b i )c i = (a 0 b 0 )c 0 + k i= (a i b i )c i > (a 0 b 0 )c + k i= (a i b i )c i = (a 0 b 0 + a b 0 )c + k i=2 (a i b i )c i > (a 0 b 0 + a b )c 2 + k i=2 (a i b i )c i = (a 0 b 0 + a b 0 + a 2 b 2 )c 2 + k i=3 (a i b i )c i. By induction, we can prove that k i=0 (a i b i )c i > (a 0 b 0 + a b + + a k b k )c k > 0. This gives the desired result. Proposition 3: Given the ( + ) elitist EA and a (λ + λ) EA (where λ 2) using global mutation and elitist selection for maximizing any unimodal function, let Φ (+) 0 = x and Φ (λ+λ) 0 = (x,, x) where x P non, then P S(λ + λ x) >. Proof: For the ( + ) EA, choose g (+) (x) to be its distance function. According to Theorem, the pointwise drift satisfies (+) (s k ) = for any s k P non. Since selection is elitist and the function is unimodal, we have (+) (x) = k Pm(s k,s l )(g (+) (s k ) g (+) (s l ))=. (2) For a (λ+λ) EA, define its distance function d (λ+λ) (X) = min{g (+) (x) : x X}. Let P k denote the set of populations whose best individual is s k (where k =, K). Provided that the parent population is X P k (where k > 0), the children population is Y and the next generation population is Z P l. Since the (µ + λ) EA adopts elitist selection and the fitness function is unimodal, the pointwise drift satisfies (µ+λ) (X)= k Z P P (X,Z)(g (+) (s k ) g (+) (s l )). l (29) Given any m < k, the children population Y and next generation population enter in the union P 0 P P m if and only if one or more parents in X is mutated into an child in the set {s 0,, s m }. Thanks to global mutation and population size λ 2, this probability is strictly larger than that of only one parent s k being mutated into the set {s 0,, s m }: m m P (X, Z) > P (s k, s l ). Z P l Since the fitness function is unimodal, we have g (+) (s k ) g (+) (s 0 ) > g (+) (s k ) g (+) (s ) > > g (+) (s k ) g (+) (s k ). Using Lemma (let a l = Z P l P (X, Z), b l = P (s k, s l ) and c l = g (+) (s k ) g (+) (s l )), we get k (λ+λ) (X) = k > Z P l P (X, Z)(g (+) (s k ) g (+) (s l )) P (s k, s l )(g (+) (s k ) g (+) (s l )) = (+) (s k ) =. Then we have (λ+λ) (X) >. Since Φ 0 P non, we have for (λ+λ) any t 0, the average drift t >. Applying Theorem 9, we get P S(λ + λ) >. Here is an explanation of the proposition. For unimodal functions, the higher the fitness of a point is, the closer to the optimal set the point is. The probability of the (λ + λ) EA (where λ 2) to generate a better individual is strictly larger than that of the ( + ) EA. Thus the expected hitting time is shortened. B. Case study : unimodal functions It is an intuitive belief that using a population can not reduce the expected running time of elitist EAs for maximizing unimodal functions. The proposition below asserts this is true for unimodal functions on the time-fitness landscape. Consider a (µ + λ) elitist EA. Mutation. Select λ individuals in Φ t and mutate them. Then generate a children population consisting of λ individuals; Elitist Selection. First select one individual with the highest fitness in Φ t Ψ t ; and then probabilistically select µ individuals from Φ t Ψ t. Proposition : Given the (+) elitist EA and a (µ+λ) EA (where µ 2 or λ 2) for maximizing a unimodal function, let Φ (+) 0 = x and Φ (µ+λ) 0 = (x,, x) where x P non, then P S(µ + λ x) λ and if λ 2, P S(µ + λ x) < λ.

9 Proof: It is sufficient to consider the case of λ >. The analysis of the (µ + ) EA is almost the same as that of Proposition 2, except two places: () without the global mutation condition, Inequality (25) is changed from < to ; (2) the conclusion is changed from P S(µ+) < to P S(µ+ ). For the ( + ) EA, choose g (+) (x) to be its distance function. According to Theorem, for any s k {s,, s K }, the pointwise drift satisfies (+) (s k ) =. Since selection is elitist and the function is unimodal, we have (+) (s k ) = k Pm(s k,s l )(g (+) (s k ) g (+) (s l ))=. (30) For the (µ + λ) EA where λ 2, let its distance function d (µ+λ) (X) = min{g (+) (x) : x X}. Let P k denote the set of populations whose best individual is s k (where k = 0,,, K). Provided that the parent population is X P k (where k > 0), the children population is Y and the next generation population is Z P l. Since the (µ + λ) EA adopts elitist selection and the fitness function is unimodal, the pointwise drift satisfies (µ+λ) (X) = k = k Z P P (X,Z)(g (+) (s k ) g (+) (s l l )) Y P l P m(x,y )(g (+) (s k ) g (+) (s l )). (3) Denote the children population Y by (s (Y ),, s λ(y ) ). Let (s (X),, s λ(x) ) be the parents from which Y are mutated. Children Y P l (where l < k) only if one or more parents is muted into s l. The probability of mutating s i(x) to s l is P m (s i(x), s l ). Since each parent is mutated independently, the probability of one or more parents is muted into s l is not more than the sum of each parents is mutated into s l. Then we have Y P l P m (X, Y ) ( g (+) (s k ) g (+) (s l ) ) λ i= P m(s i(x), s l )(g (+) (s k ) g (+) (s l )). (32) The above inequality is strict if X = (s k,, s k ). Since l < k i(x), we have and then g (+) (s l ) < g (+) (s k ) g (+) (s i(x) ), Y P l P m (X, Y ) ( g (+) (s k ) g (+) (s l ) ) λ i= P m(s i(x), s l )(g (+) (s i(x) ) g (+) (s l )). Inserting the above inequality into Equality (3), we get (µ+λ) (X) k λ i= Pm(s i(x),s l )(g (+) (s i(x) ) g (+) (s l )) = λ i= k Pm(s i(x),s l )(g (+) (s i(x) ) g (+) (s l )). (33) Since s k is the best parent in X, we have for i =,, λ, the indexes satisfy i(x) k. Then k P m(s i(x), s l )(g (+) (s i(x) ) g (+) (s l )) i(x) P m (s i(x), s l )(g (+) (s i(x) ) g (+) (s l )) =. The drift satisfies (µ+λ) (X) λ = λ. The above inequality is strict if X = (s k,, s k ). Since Φ 0 = (s k,, s k ) for some k, the average drift (µ+λ) 0 <. When t, the average drift (µ+λ) t. Applying Theorem, we obtain P S(µ + λ) < λ. The explanation of the proposition is simple. For unimodal functions, the higher the fitness of a point is, the closer to the optimal set the point is. The probability of the (µ + λ) EA to generate a better individual is not more than λ times that of the ( + ) EA. Thus the expected hitting time cannot be shortened by /λ. Example 3: Consider the TwoMax function [32] where x {0, } n. f(x) = max{ x, n x }. (3) i= A ( + λ) EA (where λ ) with elitist selection and bitwise mutation is used for maximizing the function. Bitwise Mutation. Flip each bit with a probability /n. Then generates λ children. Elitist Selection. Select the best individual from Φ t Ψ t. Table III shows that using a population reduces the expected hitting time, but increases the expected running time. TABLE III EXPERIMENTAL RESULTS FOR EXAMPLE 3 AVERAGED OVER 000 RUNS. n = 200. THE EA STARTS AT x 0 WITH x 0 = 00. population size 3 5 7 9 hitting time 2536 6 529 395 35 running time 2536 2592 265 2765 235 Let s apply Proposition to a special instance: the ( + λ) EA (using bitwise mutation and elitist selection) for maximizing the OneMax function. According to Proposition, using a population will increase the running time. The population threshold is 2. This conclusion is more accurate that that in []. The result in [] asserts that the expected running time of the ( + λ) EA is in the same order of that of the ( + ) EA by a constant factor when λ is smaller than the cut-off point. The constant could be =, > or <. But the two results are not contrary. Our result indicates that the constant factor is strictly less than when λ 2. VI. SCENARIO 3:USING A POPULATION CAN REDUCE RUNNING TIME A. Case study 5: two-paths-ii functions In the previous section, it has been proven that using a population cannot reduce the expected running time of an EA for maximizing any unimodal function on the time-fitness landscape. But this intuitive belief is not true in terms of the distance-based fitness landscape. It is demonstrated by the following case study of two-paths-ii functions.

0 Definition 7: Let S = {s 0, s,, s K+L, s K+L+ } where L < K and f a fitness function such that f(s 0 ) > f(s K+ ) > f(s K+2 ) > > f(s K+L ) >f(s ) > f(s 2 ) > > f(s K ) > f(s K+L+ ). (35) Given a ( + ) elitist EA to maximise f, f is called a two-paths-ii function to the ( + ) EA if there exist two paths to the optimum: Path (s K+L+, s K,, s, s 0 ) and Path 2 (s K+L+, s K+L,, s K+, s 0 ) such that for k =,, K and k = K + 2,, K + L, mutation probabilities P m (s k, s k ) = ; for k = K +, mutation probability P m (s k, s 0 ) = ; for k = K +L+, mutation probabilities P (s k, s K ) = p and P m (s k, s K+L ) = p where 0 < p < ; for any other i, j, P m (s i, s j ) = 0. Fig. 3 visualises the a two-paths-ii time-fitness landscape. 2) The ( + λ) EA moves from s K+L+ to s K+L. This event happens if and only if at least one mutated child is s K+L. The probability for the event happening is p λ. Calculate the pointwise drift at s K+L+ as follows: (+λ) (s K+L+ ) =( p λ )(g (+) (s K+L+ ) λg (+) (s K+L )) + p λ (g (+) (s K+L+ ) λg (+) (s K )) =( p λ )[ + pk + ( p)l λl] + p λ [ + pk + ( p)l λk] = + pk + ( p)l λl + p λ λl p λ λk = + (p p λ λ)(k L) + L( λ) >λ. (use λ < λ and (36)) (3) Calculate the pointwise drift at s {s,, s K+L }, (+λ) (s) = s S[λg (+) (s) λg (+) (s )] s 0 sk+ = (+) (s) = λ. f(x) s s K+L s K+L+ g (+) (x) s K Fig. 3. A two-paths-ii time-fitness landscape. The x axis represents the expected hitting time of the ( + ) EA. The y axis is the fitness function. Consider a ( + λ) EA where λ 2. Mutation. Mutation probabilities are identical to those in the ( + ) EA in the above definition. Elitist Selection. Select the best individual from Φ t Ψ t. Under certain condition, using a population may reduce the expected hitting time. Proposition 5: Given the ( + ) EA and a ( + λ) EA (where λ 2) for maximizing a two-paths-ii function, let Φ (+) 0 = s K+L+ and Φ (+λ) 0 = s K+L+, if the population size satisfies λ < λ where λ is given by s K λ = + (p λp λ ) K L L +, (36) then P S( + λ s K+L+ ) > λ. The scalability threshold is not less than λ. Proof: For the ( + λ) EA, define its distance function as follows: d (+λ) (x) = 0, if x = s 0, g (+) (s K+L+ ), if x = s K+L+, λg (+) (x), otherwise. (37) Consider the pointwise drift at s K+L+. There are two potential events. ) The ( + λ) EA moves from s K+L+ to s K. This event happens if and only if all mutated children are s K. The probability for the event happening is p λ. (+λ) Since Φ 0 = s K+L+, the average drift 0 > λ. For any t, Φ (+λ) t has left the point s K+L+, then average drift (+λ) t = λ. According to Theorem 9, we get that P S(+λ s K+L+ ) > λ. It is easy to understand the proposition. There are two paths to the optimum: short and long. Using a lager population, the short path is more likely to be chosen than the long path. Thus the expected hitting time is reduced. Example : Consider an instance of two-paths-ii functions. Let x {0, } n and x denote its number of -valued bits. n, if x = 0 or x = n, f(x) = x, if x θ where θ = n 2, (39) x, otherwise. The ( + λ) EA is the same as that in Example in Section IV-A. Table IV shows that using a population reduces the expected running time. TABLE IV EXPERIMENTAL RESULTS FOR EXAMPLE AVERAGED OVER 000 RUNS. n = 000. THE EA STARTS AT x 0 WITH x 0 = 99. population size 5 9 3 7 hitting time 507 2 2 2 2 running time 507 0 26 3 f is a unimodal function under Hamming distance but a twopath-ii function (multi-modal) to the ( + ) EA on the timefitness landscape. When the ( + ) EA starts at x = n 2, there are two paths towards the optima x = 0, n. The long path is x = n 2 n 3 0. The short path is x = n 2 n n. This example shows that using a population shortens the expected running time of an EA on a unimodal function in terms of the distance-based fitness landscape. It doesn t contradict Proposition, which holds for any unimodal functions on the time-fitness functions.

B. Case study 6: deceptive-like functions It is an intuitive belief that using a population may shorten the runtime of EAs on deceptive functions. This was proven for an elitist EA on a deceptive function under Hamming distance [7]. In this case study, the conclusion is generalised to deceptive-like functions in any finite set. Deceptive functions and deceptive-like functions are defined on the time-fitness landscape [32]. Definition : Let S = {s 0, s,, s K } and f a fitness function such that f(s 0 ) > f(s K ) > > f(s ). Given a ( + ) elitist EA to maximise f, f is called deceptive to the ( + ) EA if g (+) (s K ) > > g (+) (s ). According to [32], deceptive functions to a ( + ) elitist EA are the hardest among all fitness functions with the same optimum at s 0. Furthermore, according to [32], for any fitness function f(x), we can construct a ( + ) elitist EA to which f is the hardest (deceptive). Definition 9: Let S = {s 0, s,, s K } and f a fitness function such that f(s 0 ) > f(s K ) > max{f(s ),, f(s K )}. (0) Given a (+) elitist EA to maximise f, f is called deceptivelike to the ( + ) EA if g (+) (s K ) > g (+) (s k ) for any k < K. A deceptive-like time-fitness landscape is visualised in Fig.. Deceptive functions belong to deceptive-like functions. f(x) s 0 s g (+) (x) s K Fig.. A deceptive-like time-fitness landscape. The x axis is the expected hitting time of the ( + ) EA. The y axis is the fitness function. Given a fitness function f(x), consider an elitist ( + ) uses global mutation for maximising f(x). Global Mutation. Mutation probability P m (s i, s j ) > 0 for any i, j and mutation probability P m (s i, s 0 ) > P m (s K, s 0 ) for any i < K. Elitist Selection. select the best from Φ t Ψ t. The fitness function f is deceptive-like to the ( + ) EA because g (+) (s K ) > g (+) (s i ) for any i < K. This can be proven by pointwise drift analysis. Let the distance function to be d (+) (s) = g (+) (s K ) for s {s,, s K } and d (+) (s) = 0 for s = s 0. The pointwise drift satisfies (+) (s) = for s = s K and (+) (s i ) > for s i {s,, s K } due to P m (s i, s 0 ) > P m (s K, s 0 ). Hence g (+) (s K ) > g (+) (s i ) (deceptive-like). Consider a (λ+λ) EA (where λ 2) using global mutation and elitist selection with fitness diversity for maximising f(x). Global Mutation. The same as that in the ( + ) EA. Elitist Selection with Fitness Diversity. s K ) If all individuals in Φ t Ψ t have the same fitness, then choose λ individuals at random; 2) Otherwise, first select one individual with the highest fitness from Φ t Ψ t and then select one individual with a lower higher fitness. Thereafter select λ 2 individuals from Φ t Ψ t using any selection scheme. We will show that the running time of the (λ + λ) EA (where λ 2) is shorter than that of the ( + ) EA. Before the theorem and its proof, the following notation is introduced. S K def = {x S; f(x) < f(s K )}, () def P K = {X P; f(x) < f(s K ) for any x X}, (2) P m (x, S K ) def = P m (x, y), y S K (3) P m (X, P K ) def = x X P m (x, S K ). () Proposition 6: Given the (+) elitist EA and a (λ+λ) EA (where λ 2) for maximizing any deceptive-like function, if the population size satisfies λ < λ where λ is given by λ = P m(s K,S K ) min 0<i<K P m(s i,s 0) (P m(s K,S K )+min 0<i<k P m(s i,s 0))P m(s K,s 0), (5) then P S(λ+λ s K ) > λ. The scalability threshold is not less than λ. Proof: For the (λ + λ) EA, given a population X, let f(x) = max{f(x); x X} the maximal fitness of its individuals. Define the distance distance as follows: d (λ+λ) (X)= d 0 = 0, if f(x)=f(s 0), d K = g () (s K ), if f(x)=f(s K ), d K = λ if f(x)<f(s K). min 0<i<K Pm(s i,s 0 ), We calculate the pointwise drift at X = (s K,, s K ): (λ+λ) (s K,, s K ) =P (X, P opt )(d K d 0 ) + P (X, P K )(d K d K ) >P (X, P opt )(d K d K ) + P (X, P K )(d K d K ) = ( P (X, P opt ) + P (X, P K ) ) (d K d K ) >P m (s K, S K )(d K d K ) ( =P m (s K, S K ) P m (s K, s 0 ) ) λ min 0<i<K P m (s i, s 0 ) (6) >λ. (use λ < λ and (5)) (7) Calculate the pointwise drift at any X P non in which at least one of its individuals is not s K : (λ+λ) (X) =P (X, P opt )(d K d 0 ) λ max x X P m (x, s 0 ) λ. () min 0<i<k P m (s i, s 0 ) Since Φ (λ+λ) 0 = (s K,, s K ), from (7) we have the average drift 0 > λ. For any t, we know, (λ+λ) t λ. (λ+λ) According to Theorem 9, we have P S(λ + λ s K ) > λ. The proposition can be explained as follows. In the ( + ) algorithm, elitist selection cannot accept an worse solution, which is a bad strategy for deceptive-like functions. But in the