Convergence Time for Linkage Model Building in Estimation of Distribution Algorithms

Size: px
Start display at page:

Download "Convergence Time for Linkage Model Building in Estimation of Distribution Algorithms"

Transcription

1 Convergence Time for Linkage Model Building in Estimation of Distribution Algorithms Hau-Jiun Yang Tian-Li Yu TEIL Technical Report No January, 2009 Taiwan Evolutionary Intelligence Laboratory (TEIL) Department of Electrical Engineering National Taiwan University No., Sec. 4, Roosevelt Rd., Taipei, Taiwan

2 Convergence Time for Linkage Model Building in Estimation of Distribution Algorithms Hau-Jiun Yang, Tian-Li Yu National Taiwan University Abstract This paper proposes a convergence time model for linkage model building in estimation of distribution algorithms (EDAs). The model utilizes the result of population sizing for EDAs in previous work. By investigating the building-block identification rate of linkage model. We give a recurrence relation to express the proportion of identified building block in each generation. The recurrence fails to yield a closed form solution due to varying identification rate. Therefore, we derive upper and lower bounds instead by assuming rapid allelic convergence and fixed identification rate respectively. Specifically, The linkage model convergence time is bounded by Ω(ln(m)) and O(m), where m is the number of building blocks. Empirically, experiment results agree with the proposed bounds. INTRODUCTION Estimation of distribution algorithms (EDAs) (Larrañaga, Lozano, & eds, 2002) have lots of attention in recent decades. It has been proposed as a strong extension of genetic algorithms. EDAs solve deceptive and nonseparable functions that genetic algorithms have difficulties to optimize by learning linkage (Mhlenbein & Hns, 2006). Linkage learning is essential to EDA success, because it can conserve building blocks under recombination. Many methods (Goldberg, 2002) have been developed to solve the so-called linkage-learning problem, and resulted in many different EDAs, including ecga (Harik, 999), BOA (Pelikan & Goldberg, 999), COMIT (Baluja & Davies, 997), UMDA (Mühlenbein & Paass, 996), MIMIC (Bonet, Isbell, & Viola, 997), D 5 (Tsuji, Munetomo, & Akama, 2004), EBNA (Etxeberria & Larrañaga, 999) and DSMGA (Yu, Goldberg, Yassine, & Chen, 2003). To solve problems with EDAs, we need to understand the race between alleles and linkages (Goldberg, 998; Goldberg, 2002). Allelic convergence means a superior individual taking over the population. Linkage convergence means that all of the building block are identified which ensures the superior individual never be recombined and makes the answer to the question comes convincingly. Here comes a question: Whether should alleles converge faster than linkages or not? The condition where the linkages convergence time is less than or equal to the takeover time is most advantageous, because prior to the best alleles dominating the population, linkage should be created to innovate and recombine a better individual. Therafter this better individual starts to dominate the population. The convergence time of simple genetic algorithm (sga) has been modeled under different selection schemes (Thierens & Goldberg, 994). The issue of convergence time is equally critical in EDAs. Once linkage converge to the best combination of gene, alleles will be hard to converge in wrong solution. So linkage convergence is more important than alleles convergence in EDAs. The purpose of this paper is to develop a convergence time model for linkage model building in EDAs. The model utilizes the population-sizing model of EDAs in previous work (Yu, Sastry, Goldberg, & Pelikan, 2007). First, this paper derives the linkage model identification rates of EDAs at the first generation from prior research. Then observing the reason of the changing model identification rate and some approximation will be used to approach the phenomena and then we give a recurrence relation to express the proportion of the identified building block. This paper then shows the derivation of the bound by the proposed recurrence relation function which

3 is hard to derive to a closed form. Finally, experiments are conducted to verify the linkdage time convergence model. 2 Convergence Time Time to convergence model in sga have been proposed. Since our methodology is depends on EDAs, we outline the models in sga conducted by takeover time and convergence model briefly. 2. Takeover Time At the beginning of GA, convergence time is characterized with takeover time (Goldberg & Deb, 99). The takeover time is compared by a number of selection schemes commonly used in sga. The comparison is given by selection scheme takeover time Proportionate Selection c n ln n Linear ranking c 0 = 2 log n + log(ln n) Tournament s ln s [ln n + ln(ln n)] Genitor c 0 ln(n ) where n is population size, and s is the selection pressure. 2.2 Time Convergence model Thierens (Thierens & Goldberg, 994) proposed a model by assuming a normally distributed fitness function and predict the proportion of optimal alleles in function of the number of generations. Different selection schemes are derived that allow us to compare the GA selection schemes in terms of their convergence characteristics. The comparison is given by selection scheme Convergence time(g conv ) Proportionate l ln(2ɛ) π Tournament s 2 πl π Truncation 2. l i π Elitist Recombination 2 πl where l is the problem size, i is selection intensity (Blickle & Thiele, 995), ɛ is a failure probability. 3 Linkage Model identification rate for EDAs Linkage model identification rate means the proportion of the building blocks (BBs) will be identified within the unidentified BBs for next generation. In this section, we investigate the linkage model identification rate of EDAs at the first generation. If we want to identify all BBs at the end of EDA, sufficient population size is needed. If population is too small, alleles will converge faster than linkages, so as to we can not convince the solution is correct. The following derivation adopts a decomposable problem composed of bipolar Royal road function. The Royal road function serves as a worst case scenario for model building because given the minimal fitness difference d min, the fitness differences between the best schema and all other (2 k ) schemata are all d min. A bipolar Royal road function of order k is defined as follows to equalize the growth rates of 0 and for every gene and the total number of BBs is proportional to the problem size. R k ( x) =, if x =... }{{} k, if x = }{{} k d, otherwise. () 2

4 3.5 x x Figure : The difference between error function (erf) and its approximation (erf (x)). The difference is shown by erf (x) erf(x) erf(x). The maximum error happens when x is close to.3 The fitness of an additively decomposable problem with m BBs is defined as f( x) = m i=0 R k (x ik+ x ik+2...x ik+k ). (2) The means and the variances of the sampled mutual information for finite population was derived in (Yu, Sastry, Goldberg, & Pelikan, 2007). Using these model, we can derive linkage model identification rate of EDAs at the first generation. The distribution of sampled mutual information can be approximated as a Gaussian distribution(hutter & Zaffalon, 2005). The decision-making error can be calculated as follows. Define a variable τ as τ = E[Z] ln s to.6 c m n (3) V ar[z] sto d where m = 2 k πmσ BB where c is constant, s to is selection pressure with tournament selection, n is population size, m is number of BB, k is the order of each BB and σbb 2 is the fitness variance of a BB The decision error ɛ is Φ(τ) and one of the approximation of error function is given by erf 2 (x) e 8(π 3) where a = 3π(π 4) 4/π+ax 2 x2 +ax 2 (4) The difference between error function and its approximation is shown in Figure. The approximation have maximum error when x is close to.3 and the error have little influence to our work. π 4 4/π+ax 2 +ax 2 only influence the result when x is close to zero, it make error of the approximation is at most. So we can approximate it to and calculate the error function as follow. erf(x) By the approximation of error function, we can calculate ɛ as e x2 (5) ɛ = Φ(τ) = 2 ( + erf( τ )) (6) 2 2 e τ2 2 (7) 2 3

5 For large m, τ is small and the decision error can be simplified as ɛ τ2 e 2 (8) 4 For a problem with m BBs, each BB is order k. There are k bits in the BB should be identified and the BB should be identified in m BBs. So the identification rate of the initial state Q 0 of EDAs should be ( ɛ) km Q 0 (9) Usually, Q 0 is not close to zero when we provide sufficient population size that make all the BBs be identified. With a not too small Q 0 and large m, we can infer the value of ɛ is small from equation (9) easily. So equation (9) can be approximated as Q 0 kmɛ = kme c (ln s to.6 )2 2 m n 2 s to = kme cn m (0) where c = c 2 (ln s to.6 )2 d 2 2 2k s to πσ 2 BB proportion of BB (m) population size(n) Figure 2: n vs Q 0 graph fixed m. Given a small population size, the quality will become zero, but given sufficient population size, the trend of line will grow rapidly. Figure 2 shows the experimental result for ecga on an (m, k)-trap problem(goldberg, 989), where m = 40 and k = 3. When we supply sufficient population size, the trend of Q 0 grows rapidly. In this paper, we only consider the case when we supply sufficient population size. So the experiment result is similar to equation (0). 4 TIME TO MODEL CONVERGENCE Previous section derived the linkage model identification rate of EDAs at first generation. This section we will discuss the identification rate in the underway EDAs. Define Q t is the ratio of identified BBs at t generation. β t and α t are the linkage model identification rate and linkage model destruction rate at the t generation respectively. The destruction rate means the proportion of identified model will be destroyed in the next generation. 4

6 4. The Changing Linkage Model Identification Rate With empirical result, we know that the model identification rate β is increasing though the running time (Figure 3). We start to analyze the reason of the increasing β. proportion of BB (m) Q t β t α t generation (t) (a) bipolar Royal road function proportion of BB (m) Q t β t α t generation (t) (b) Trap function Figure 3: The effect of α is too small to neglect, β is increasing when we supply enough population size. Trap function need a larger population parameter to ensure the entire model identified. All the EDA methods share a similar high level algorithm presented as below (Shan, McKay, Essam, & Abbass, 2006):. Generate a population P randomly. 2. Select a set of fitter individuals G from P 3. Estimate a probabilistic model M over G 4. Sample from the probabilistic model M to obtain a set of new individuals G 5. Incorporate G into population P 6. If the termination condition is not satisfied, go to 2. EDAs start with random initialize and execute from step 2 to step 6 circularly. The variance of fitness between superior and inferior individual larger can help the linkage model better. In the other word, the more superior individual generate, the better linkage model we have. Remove step 3 and modify step 4 to random shuffle in each gene that make EDAs work without linkage learning. Execute EDAs for bipolar Royal road function without learning linkage, the ratio of 5

7 0s and s in each gene is fixed in 0.5. Without learning linkage, we can discover the number of superior individual copies will increase slowly. So the key to cause the number of superior individuals climbing is in the process of learning linkage. When EDA is learning linkages, EDA will check all possible combination of gene, and EDA will adopt the combination which have the highest entropy. Once a correct BB is identified, the gene in the BB will converge to the best solution of the model. When the best alleles dominate the population in the BB, the BB will have maximum fitness which supply the maximum entropy to the entire linkage model. To other BBs, the identified BBs provide no external noise to them. So we can consider the noise come from the identified BBs is zero. Then we show the progress that follow the step in (Yu, Sastry, Goldberg, & Pelikan, 2007). Define the follow notation for schemata: H xy = xy... }{{} l 2 () where x and y are 0 or, and l is the problem size. Define two sets H + = {H 00, H } and H = {H 0, H 0 }, and let F + and F be their corresponding fitness values: F + = f(h + ) = f(h 00 ) + f(h ) (2) F = f(h ) = f(h 0 ) + f(h 0 ) (3) According to the central limit theorem, the distribution of F + and F can be approximated as Gaussian distributions when population size is large. The variances of F + and F, defined as σ 2 F + and σ 2 F respectively, are different but very close. By treating other (m ) BBs as external noises, these variances can be bounded and approximated as: (m )σ 2 BB σ 2 F + mσ 2 BB (m )σ 2 BB σ 2 F mσ 2 BB σf 2 + σf 2 = mσbb( 2 O( )) (4) m where σbb 2 is the fitness variance of a BB. The difference between those two variances is small and can be neglected when m is large. Define Z = F + F. Z is a normal distributed random variable with the following mean and variance. E[Z] = d 2 k 2 (5) V ar[z] = σ 2 F + + σ 2 F = 2mσ 2 BB( O( m ) (6) Once a BB is identified, the best alleles of the BB will never be destroyed. The expected value of the best alleles copies is n 2 k, because the ratio of the 0 and for every gene are equal in bipolar Royal road function in each generation, where n is population size. The taking over time t of the best alleles can be calculated as: n 2 t = n 2 k t = k (7) the take up time t is equal to k, k is not too large usually. So we approximate t to. That means once BB is identified, the alleles in the BB will takeover the population immediately. In the other word, the BB will have no noisy to the linkage identification immediately. So, the means and variances are changed and shown as bellow. E[Z] = d 2 k 2 (8) V ar[z] = 2(m )σbb( 2 O( m ) (9) 6

8 Extend to m( Q t ) BBs has been identified in generaion t. The means and variances of Z at generation t can be calculated as: E[Z] = d 2 k 2 (20) V ar[z] = 2m( Q t )σbb( 2 O( m( Q t ) ) (2) Follow the step in (Yu, Sastry, Goldberg, & Pelikan, 2007), we can derive the result finally as bellow. β t = km( Q t )e cn m( Q t ) (22) 4.2 The Recurrence relation to the identified building block ratio Q t+ is depend on α t and β t and Q t. The recurrence formula can be derived easily as follow by combining destruction ratio ( α t ) Q t and identification ratio β t ( Q t ), and the initial state Q 0 has been derived in previous section (Equation 0). Q t+ = ( α t )Q t + β t ( Q t ) (23) This recurrence relation can provide exact Q t+ by providing exact α t, β and Q t. According to the experimental result (Figure 3), α t is always zero in bipolar Royal Road function. In (m, k)-trap function (Deb & Goldbeg, 99;?), α is small, too. At the beginning of EDAs, Q t is often small, so the destruction ratio ( α t )Q t can be negligible. So we assume α to 0 and the recurrence function can be simplified to Q t+ = Q t + β t ( Q t ) (24) 4.3 The upper bound of linkage model convergence time With some simple transposition and substitution of equation (24), the equation will be cn Q t+ Q t = ( km( Q t )e m( Q t ) )( Qt ) (25) Approximating the difference equation with the corresponding differential equation dq t dt dt = = ( km( Q t )e cn m( Q t ) )( Qt ) (26) dq ( km( Q t )e cn t (27) m( Q t ) )( Qt ) Integrate the both side of the above equation, the right side of the equal mark is hard to integrate to a closed form straightly. So we make an approximation that plus Q 2 t Q t to the denominator. Q 2 t Q t is smaller than zero, because the value of Q t is between 0 and. And we can integrate the equation as follow. cn t Qt Q 0 ( kme cn m( Q t ) )( Qt ) 2 dq t (28) let P t = m( Q t ), then dq t = m( Qt)2 c n dp t m t c n kme P dp t (29) t m e P t = c n e P dp t t (30) km m ] = [ln(e cn Qt m( Q t ) km) + C (3) c n Q 0 = m c n ln e e cn m( Q t ) cn m( Q 0) 7 km km (32)

9 We obtain a time convergence model expressing the proportion Q t in function of the number of generations t. c n Q t (33) m ln(e nt c m (e cn m( Q 0) km) + km) To calculate the linkage convergence time, we let the proportion Q t come close to m. Q t t m c n ln c n m ln(e nt c m (e cn m( Q 0) km) + km) e cn km e cn m( Q 0) km m (34) From initial state Q 0 in equation (0), we can replace n by n = m c ln km Q 0 t = ln km Q 0 ln km Q 0 ln e m ln km Q 0 km ln e ln Q km 0 ( Q 0 ) km ( km Q 0 ) m km ( km ( Q Q 0 ) 0) km (35) (36) For large m, km can be neglect then inequality 36 can be simplified as t m Q 0 (37) 4.4 The lower bound of linkage model convergence time In our approximation, once BB identified, the noise come from the BB become zero immediately. To avoid overestimation in the approximation, we give the lower bound roughly. We consider the case that β t never increase when EDAs is working. In the other words, β t is fixed in Q 0. The equation can describe as Q t = Q t + Q 0 ( Q t ) (38) = ( Q 0 ) t+ (39) Let the proportion Q t come close to m and then we can obtain convergence time t Q t = ( Q 0 ) t+ m t + ln m ln( Q 0 ) (40) (4) ln( Q 0 ) can be approximated as Q 0 when 0 ( Q 0 ) 2. Finally, the lower bound of m t relation is 5 Experimental Result and discussion t + lnm Q 0 (42) The relation between number of building block and linkage model convergence time is presented in Figure 4. In equation (37) and (42), we know the linkage model convergence time t can be bounded by O(m) and Ω(ln m). Figure 4(a) shows the experiment with bipolar royal road function where Q 0 is 0. and selection pressure is 4 and figure 4(b) shows the experiment with trap function where Q 0 is 0.25 and selection pressure is 6. When m is small, the trend is almost scaled as linear. Actually, the external noise mentioned in previous section is not zero and alleles does not takeover 8

10 convergence time(generation) ecga, k=3, Q=0. O(m / Q 0 ) O(ln(m)) number of BBs (a) bipolar Royal road function convergence time(generation) ecga, k=3, Q=0.25 O(m / Q 0 ) O(ln(m)) number of BBs (b) Trap function Figure 4: The relation between the problem size and convergence time. The convergence time t is bounded by O(m) and Ω(ln m) 9

11 the population immediately when BB identified. These are the reason to explain why the upper bound is not predict accurately. In this paper, we does not discuss the condition that BBs are not fully identified. An example to the condition, a build block is defined as -2-3, but EDAs only identify -2 as a building block. In our proposed model, superior alleles start to grow only full BB is identified. When the order of BB is very large, EDAs are hard to identify the full BB directly, and the superior alleles will never increase without learing linkage. Therefore, the order of BB is smaller, the upper bound of our proposed model is more accurate. 0.9 model allele proportion generation(t) Figure 5: The comparison between alleles and linkage. Figure 5 shows the comparison between alleles and linkages on trap function where m = CONCLUSION In this paper we present a model that bound the linkage model convergence time for EDAs. By observing the EDAs flow, we discover the external noise decrease that make the BB identified rate increase. Furthermore, we give a recurrence relation to express the proportion of identified building block. The upper bound is based on an approximation that the noise comes from a BB, which is identified in a previous generation, becomes zero. Because it is difficult to give a closed form to the recurrence relation, we take some approximate in the derivation. And we give a lower bound to the convergence time assuming the BB identification rate is always fixed in Q 0. Our model presents the required linkage model convergence time t is bounded by Θ(ln m) and Θ(m). Our proposed model is scaled as Θ(m) originally. With the grow of problem size, the error grows, too. Such the trend is similar to Θ(ln m). Experimental results are show good agreement with the bound. References Baluja, S., & Davies, S. (997). Combining multiple optimization runs with optimal dependency trees (Technical Report). Blickle, T., & Thiele, L. (995). A comparison of selection schemes used in genetic algorithms. Bonet, J. S. D., Isbell, C. L., & Viola, P. (997). Mimic: Finding optima by estimating probability densities. In Advances in Neural Information Processing Systems (pp. 424). The MIT Press. Deb, K., & Goldbeg, D. E. (99). Analyzing deception in trap functions. IlliGAL Report No. 9009, University of Illinois at Urbana-Chmapaign, Illinois Genetic Algorithms Laboratory, Urbana, IL. Etxeberria, R., & Larrañaga, P. (999). Global optimization using bayesian networks. Second Symposium on Artificial Intelligence (CIMAF-99), Goldberg, D. E. (989). Simple genetic algorithms and the minimal, deceptive problem. Genetic Algorithms and Simulated Annealing, Goldberg, D. E. (998). The race, the hurdle, and the sweet spot: Lessons from genetic algorithms for the automation of design innovation and creativity. IlliGAL Report No , University of Illinois at Urbana-Chmapaign, Illinois Genetic Algorithms Laboratory, Urbana, IL. 0

12 Goldberg, D. E. (2002). The design of innovation: Lessons from and for competent genetic algorithms. Norwell, MA, USA. Goldberg, D. E., & Deb, K. (99). A comparative analysis of selection schemes used in genetic algorithms. Harik, G. R. (999). Linkage learning via probabilistic modeling in the ecga (Technical Report). Hutter, M., & Zaffalon, M. (2005). Distribution of mutual information from complete and incomplete data. Computational Statistics & Data Analysis, 48 (3), to appear. Larrañaga, P., Lozano, J. A., & eds (2002). Estimation of distribution algorithms: A new tool for evolutionary computation. Kluwer Academic Publishers, Boston, MA. Mühlenbein, H., & Paass, G. (996). From recombination of genes to the estimation of distributions i. binary parameters. In PPSN IV: Proceedings of the 4th International Conference on Parallel Problem Solving from Nature (pp ). London, UK: Springer-Verlag. Mhlenbein, H., & Hns, R. (2006). Scalable optimization via probabilistic modeling: The factorized distribution algorithm and the minimum relative entropy principle. Springer Berlin / Heidelberg. Pelikan, M., & Goldberg, D. E. (999). Boa: The bayesian optimization algorithm. (pp ). Morgan Kaufmann. Shan, Y., McKay, R. I., Essam, D., & Abbass, H. A. (2006). Scalable optimization via probabilistic modeling: A survey of probabilistic model building genetic programming. Springer Berlin / Heidelberg. Thierens, D., & Goldberg, D. E. (994). Convergence models of genetic algorithm selection schemes. In PPSN III: Proceedings of the International Conference on Evolutionary Computation. The Third Conference on Parallel Problem Solving from Nature (pp. 9 29). London, UK: Springer-Verlag. Tsuji, M., Munetomo, M., & Akama, K. (2004). Modeling dependencies of loci with string classification according to fitness differences. Genetic and Evolutionary Computation GECCO 2004, Yu, T.-L., Goldberg, D. E., Yassine, A., & Chen, Y.-P. (2003). Genetic algorithm design inspired by organizational theory: Pilot study of a dependency structure matrix driven genetic algorithm. Proceedings of Artificial Neural Networks in Engineering Yu, T.-L., Sastry, K., Goldberg, D. E., & Pelikan, M. (2007). Population sizing for entropy-based model building in discrete estimation of distribution algorithms. In GECCO 07: Proceedings of the 9th annual conference on Genetic and evolutionary computation (pp ). New York, NY, USA: ACM.

Interaction-Detection Metric with Differential Mutual Complement for Dependency Structure Matrix Genetic Algorithm

Interaction-Detection Metric with Differential Mutual Complement for Dependency Structure Matrix Genetic Algorithm Interaction-Detection Metric with Differential Mutual Complement for Dependency Structure Matrix Genetic Algorithm Kai-Chun Fan Jui-Ting Lee Tian-Li Yu Tsung-Yu Ho TEIL Technical Report No. 2010005 April,

More information

Spurious Dependencies and EDA Scalability

Spurious Dependencies and EDA Scalability Spurious Dependencies and EDA Scalability Elizabeth Radetic, Martin Pelikan MEDAL Report No. 2010002 January 2010 Abstract Numerous studies have shown that advanced estimation of distribution algorithms

More information

Estimation-of-Distribution Algorithms. Discrete Domain.

Estimation-of-Distribution Algorithms. Discrete Domain. Estimation-of-Distribution Algorithms. Discrete Domain. Petr Pošík Introduction to EDAs 2 Genetic Algorithms and Epistasis.....................................................................................

More information

Scalability of Selectorecombinative Genetic Algorithms for Problems with Tight Linkage

Scalability of Selectorecombinative Genetic Algorithms for Problems with Tight Linkage Scalability of Selectorecombinative Genetic Algorithms for Problems with Tight Linkage Kumara Sastry 1,2 and David E. Goldberg 1,3 1 Illinois Genetic Algorithms Laboratory (IlliGAL) 2 Department of Material

More information

State of the art in genetic algorithms' research

State of the art in genetic algorithms' research State of the art in genetic algorithms' Genetic Algorithms research Prabhas Chongstitvatana Department of computer engineering Chulalongkorn university A class of probabilistic search, inspired by natural

More information

Improved Runtime Bounds for the Univariate Marginal Distribution Algorithm via Anti-Concentration

Improved Runtime Bounds for the Univariate Marginal Distribution Algorithm via Anti-Concentration Improved Runtime Bounds for the Univariate Marginal Distribution Algorithm via Anti-Concentration Phan Trung Hai Nguyen June 26, 2017 School of Computer Science University of Birmingham Birmingham B15

More information

Analyzing Probabilistic Models in Hierarchical BOA on Traps and Spin Glasses

Analyzing Probabilistic Models in Hierarchical BOA on Traps and Spin Glasses Analyzing Probabilistic Models in Hierarchical BOA on Traps and Spin Glasses Mark Hauschild, Martin Pelikan, Claudio F. Lima, and Kumara Sastry IlliGAL Report No. 2007008 January 2007 Illinois Genetic

More information

Analysis of Epistasis Correlation on NK Landscapes with Nearest-Neighbor Interactions

Analysis of Epistasis Correlation on NK Landscapes with Nearest-Neighbor Interactions Analysis of Epistasis Correlation on NK Landscapes with Nearest-Neighbor Interactions Martin Pelikan MEDAL Report No. 20102 February 2011 Abstract Epistasis correlation is a measure that estimates the

More information

Initial-Population Bias in the Univariate Estimation of Distribution Algorithm

Initial-Population Bias in the Univariate Estimation of Distribution Algorithm Initial-Population Bias in the Univariate Estimation of Distribution Algorithm Martin Pelikan and Kumara Sastry MEDAL Report No. 9 January 9 Abstract This paper analyzes the effects of an initial-population

More information

Gaussian EDA and Truncation Selection: Setting Limits for Sustainable Progress

Gaussian EDA and Truncation Selection: Setting Limits for Sustainable Progress Gaussian EDA and Truncation Selection: Setting Limits for Sustainable Progress Petr Pošík Czech Technical University, Faculty of Electrical Engineering, Department of Cybernetics Technická, 66 7 Prague

More information

Analyzing Probabilistic Models in Hierarchical BOA on Traps and Spin Glasses

Analyzing Probabilistic Models in Hierarchical BOA on Traps and Spin Glasses Analyzing Probabilistic Models in Hierarchical BOA on Traps and Spin Glasses Mark Hauschild Missouri Estimation of Distribution Algorithms Laboratory (MEDAL) Dept. of Math and Computer Science, 320 CCB

More information

Behaviour of the UMDA c algorithm with truncation selection on monotone functions

Behaviour of the UMDA c algorithm with truncation selection on monotone functions Mannheim Business School Dept. of Logistics Technical Report 01/2005 Behaviour of the UMDA c algorithm with truncation selection on monotone functions Jörn Grahl, Stefan Minner, Franz Rothlauf Technical

More information

Expanding From Discrete To Continuous Estimation Of Distribution Algorithms: The IDEA

Expanding From Discrete To Continuous Estimation Of Distribution Algorithms: The IDEA Expanding From Discrete To Continuous Estimation Of Distribution Algorithms: The IDEA Peter A.N. Bosman and Dirk Thierens Department of Computer Science, Utrecht University, P.O. Box 80.089, 3508 TB Utrecht,

More information

Upper Bounds on the Time and Space Complexity of Optimizing Additively Separable Functions

Upper Bounds on the Time and Space Complexity of Optimizing Additively Separable Functions Upper Bounds on the Time and Space Complexity of Optimizing Additively Separable Functions Matthew J. Streeter Computer Science Department and Center for the Neural Basis of Cognition Carnegie Mellon University

More information

Fitness Inheritance in Multi-Objective Optimization

Fitness Inheritance in Multi-Objective Optimization Fitness Inheritance in Multi-Objective Optimization Jian-Hung Chen David E. Goldberg Shinn-Ying Ho Kumara Sastry IlliGAL Report No. 2002017 June, 2002 Illinois Genetic Algorithms Laboratory (IlliGAL) Department

More information

From Mating Pool Distributions to Model Overfitting

From Mating Pool Distributions to Model Overfitting From Mating Pool Distributions to Model Overfitting Claudio F. Lima DEEI-FCT University of Algarve Campus de Gambelas 85-39, Faro, Portugal clima.research@gmail.com Fernando G. Lobo DEEI-FCT University

More information

New Epistasis Measures for Detecting Independently Optimizable Partitions of Variables

New Epistasis Measures for Detecting Independently Optimizable Partitions of Variables New Epistasis Measures for Detecting Independently Optimizable Partitions of Variables Dong-Il Seo, Sung-Soon Choi, and Byung-Ro Moon School of Computer Science & Engineering, Seoul National University

More information

Probabilistic Model-Building Genetic Algorithms

Probabilistic Model-Building Genetic Algorithms Probabilistic Model-Building Genetic Algorithms Martin Pelikan Dept. of Math. and Computer Science University of Missouri at St. Louis St. Louis, Missouri pelikan@cs.umsl.edu Foreword! Motivation Genetic

More information

Program Evolution by Integrating EDP and GP

Program Evolution by Integrating EDP and GP Program Evolution by Integrating EDP and GP Kohsuke Yanai and Hitoshi Iba Dept. of Frontier Informatics, Graduate School of Frontier Science, The University of Tokyo. 7-3-1 Hongo, Bunkyo-ku, Tokyo 113-8654,

More information

Linkage Identification Based on Epistasis Measures to Realize Efficient Genetic Algorithms

Linkage Identification Based on Epistasis Measures to Realize Efficient Genetic Algorithms Linkage Identification Based on Epistasis Measures to Realize Efficient Genetic Algorithms Masaharu Munetomo Center for Information and Multimedia Studies, Hokkaido University, North 11, West 5, Sapporo

More information

Probabilistic Model-Building Genetic Algorithms

Probabilistic Model-Building Genetic Algorithms Probabilistic Model-Building Genetic Algorithms a.k.a. Estimation of Distribution Algorithms a.k.a. Iterated Density Estimation Algorithms Martin Pelikan Foreword Motivation Genetic and evolutionary computation

More information

Gene Pool Recombination in Genetic Algorithms

Gene Pool Recombination in Genetic Algorithms Gene Pool Recombination in Genetic Algorithms Heinz Mühlenbein GMD 53754 St. Augustin Germany muehlenbein@gmd.de Hans-Michael Voigt T.U. Berlin 13355 Berlin Germany voigt@fb10.tu-berlin.de Abstract: A

More information

about the problem can be used in order to enhance the estimation and subsequently improve convergence. New strings are generated according to the join

about the problem can be used in order to enhance the estimation and subsequently improve convergence. New strings are generated according to the join Bayesian Optimization Algorithm, Population Sizing, and Time to Convergence Martin Pelikan Dept. of General Engineering and Dept. of Computer Science University of Illinois, Urbana, IL 61801 pelikan@illigal.ge.uiuc.edu

More information

From Mating Pool Distributions to Model Overfitting

From Mating Pool Distributions to Model Overfitting From Mating Pool Distributions to Model Overfitting Claudio F. Lima 1 Fernando G. Lobo 1 Martin Pelikan 2 1 Department of Electronics and Computer Science Engineering University of Algarve, Portugal 2

More information

Hierarchical BOA Solves Ising Spin Glasses and MAXSAT

Hierarchical BOA Solves Ising Spin Glasses and MAXSAT Hierarchical BOA Solves Ising Spin Glasses and MAXSAT Martin Pelikan 1,2 and David E. Goldberg 2 1 Computational Laboratory (Colab) Swiss Federal Institute of Technology (ETH) Hirschengraben 84 8092 Zürich,

More information

Bistability of the Needle Function in the Presence of Truncation Selection

Bistability of the Needle Function in the Presence of Truncation Selection Bistability of the Needle Function in the Presence of Truncation Selection Alden Wright 1 and Greg Cripe 2 1 University of Montana, Missoula MT USA 2 Spokane Falls Community College, Spokane WA USA Abstract.

More information

V. Evolutionary Computing. Read Flake, ch. 20. Assumptions. Genetic Algorithms. Fitness-Biased Selection. Outline of Simplified GA

V. Evolutionary Computing. Read Flake, ch. 20. Assumptions. Genetic Algorithms. Fitness-Biased Selection. Outline of Simplified GA Part 5A: Genetic Algorithms V. Evolutionary Computing A. Genetic Algorithms Read Flake, ch. 20 1 2 Genetic Algorithms Developed by John Holland in 60s Did not become popular until late 80s A simplified

More information

V. Evolutionary Computing. Read Flake, ch. 20. Genetic Algorithms. Part 5A: Genetic Algorithms 4/10/17. A. Genetic Algorithms

V. Evolutionary Computing. Read Flake, ch. 20. Genetic Algorithms. Part 5A: Genetic Algorithms 4/10/17. A. Genetic Algorithms V. Evolutionary Computing A. Genetic Algorithms 4/10/17 1 Read Flake, ch. 20 4/10/17 2 Genetic Algorithms Developed by John Holland in 60s Did not become popular until late 80s A simplified model of genetics

More information

IV. Evolutionary Computing. Read Flake, ch. 20. Assumptions. Genetic Algorithms. Fitness-Biased Selection. Outline of Simplified GA

IV. Evolutionary Computing. Read Flake, ch. 20. Assumptions. Genetic Algorithms. Fitness-Biased Selection. Outline of Simplified GA IV. Evolutionary Computing A. Genetic Algorithms Read Flake, ch. 20 2014/2/26 1 2014/2/26 2 Genetic Algorithms Developed by John Holland in 60s Did not become popular until late 80s A simplified model

More information

Probabilistic Model-Building Genetic Algorithms

Probabilistic Model-Building Genetic Algorithms Probabilistic Model-Building Genetic Algorithms a.k.a. Estimation of Distribution Algorithms a.k.a. Iterated Density Estimation Algorithms Martin Pelikan Missouri Estimation of Distribution Algorithms

More information

Truncation Selection and Gaussian EDA: Bounds for Sustainable Progress in High-Dimensional Spaces

Truncation Selection and Gaussian EDA: Bounds for Sustainable Progress in High-Dimensional Spaces Truncation Selection and Gaussian EDA: Bounds for Sustainable Progress in High-Dimensional Spaces Petr Pošík Czech Technical University in Prague Faculty of Electrical Engineering, Department of Cybernetics

More information

Variable Dependence Interaction and Multi-objective Optimisation

Variable Dependence Interaction and Multi-objective Optimisation Variable Dependence Interaction and Multi-objective Optimisation Ashutosh Tiwari and Rajkumar Roy Department of Enterprise Integration, School of Industrial and Manufacturing Science, Cranfield University,

More information

Bayesian Optimization Algorithm, Population Sizing, and Time to Convergence

Bayesian Optimization Algorithm, Population Sizing, and Time to Convergence Preprint UCRL-JC-137172 Bayesian Optimization Algorithm, Population Sizing, and Time to Convergence M. Pelikan, D. E. Goldberg and E. Cantu-Paz This article was submitted to Genetic and Evolutionary Computation

More information

A Statistical Genetic Algorithm

A Statistical Genetic Algorithm A Statistical Genetic Algorithm Angel Kuri M. akm@pollux.cic.ipn.mx Centro de Investigación en Computación Instituto Politécnico Nacional Zacatenco México 07738, D.F. Abstract A Genetic Algorithm which

More information

Hierarchical BOA, Cluster Exact Approximation, and Ising Spin Glasses

Hierarchical BOA, Cluster Exact Approximation, and Ising Spin Glasses Hierarchical BOA, Cluster Exact Approximation, and Ising Spin Glasses Martin Pelikan and Alexander K. Hartmann MEDAL Report No. 2006002 March 2006 Abstract This paper analyzes the hierarchical Bayesian

More information

WITH. P. Larran aga. Computational Intelligence Group Artificial Intelligence Department Technical University of Madrid

WITH. P. Larran aga. Computational Intelligence Group Artificial Intelligence Department Technical University of Madrid Introduction EDAs Our Proposal Results Conclusions M ULTI - OBJECTIVE O PTIMIZATION WITH E STIMATION OF D ISTRIBUTION A LGORITHMS Pedro Larran aga Computational Intelligence Group Artificial Intelligence

More information

Bounding the Population Size in XCS to Ensure Reproductive Opportunities

Bounding the Population Size in XCS to Ensure Reproductive Opportunities Bounding the Population Size in XCS to Ensure Reproductive Opportunities Martin V. Butz and David E. Goldberg Illinois Genetic Algorithms Laboratory (IlliGAL) University of Illinois at Urbana-Champaign

More information

Evolutionary Computation

Evolutionary Computation Evolutionary Computation - Computational procedures patterned after biological evolution. - Search procedure that probabilistically applies search operators to set of points in the search space. - Lamarck

More information

Hierarchical BOA, Cluster Exact Approximation, and Ising Spin Glasses

Hierarchical BOA, Cluster Exact Approximation, and Ising Spin Glasses Hierarchical BOA, Cluster Exact Approximation, and Ising Spin Glasses Martin Pelikan 1, Alexander K. Hartmann 2, and Kumara Sastry 1 Dept. of Math and Computer Science, 320 CCB University of Missouri at

More information

Program Evolution by Integrating EDP and GP

Program Evolution by Integrating EDP and GP Program Evolution by Integrating EDP and GP Kohsuke Yanai, Hitoshi Iba Dept. of Frontier Informatics, Graduate School of Frontier Science, The University of Tokyo. 7-3-1 Hongo, Bunkyo-ku, Tokyo 113-8654,

More information

Signal, noise, and genetic algorithms

Signal, noise, and genetic algorithms Oregon Health & Science University OHSU Digital Commons CSETech June 1991 Signal, noise, and genetic algorithms Mike Rudnick David E. Goldberg Follow this and additional works at: http://digitalcommons.ohsu.edu/csetech

More information

Performance of Evolutionary Algorithms on NK Landscapes with Nearest Neighbor Interactions and Tunable Overlap

Performance of Evolutionary Algorithms on NK Landscapes with Nearest Neighbor Interactions and Tunable Overlap Performance of Evolutionary Algorithms on NK Landscapes with Nearest Neighbor Interactions and Tunable Overlap Martin Pelikan, Kumara Sastry, David E. Goldberg, Martin V. Butz, and Mark Hauschild Missouri

More information

c Copyright by Martin Pelikan, 2002

c Copyright by Martin Pelikan, 2002 c Copyright by Martin Pelikan, 2002 BAYESIAN OPTIMIZATION ALGORITHM: FROM SINGLE LEVEL TO HIERARCHY BY MARTIN PELIKAN DIPL., Comenius University, 1998 THESIS Submitted in partial fulfillment of the requirements

More information

An Effective Chromosome Representation for Evolving Flexible Job Shop Schedules

An Effective Chromosome Representation for Evolving Flexible Job Shop Schedules An Effective Chromosome Representation for Evolving Flexible Job Shop Schedules Joc Cing Tay and Djoko Wibowo Intelligent Systems Lab Nanyang Technological University asjctay@ntuedusg Abstract As the Flexible

More information

STATISTICAL OPTIMIZATION OF COMPOSITE LAMINATES: INTRODUCTION OF COUPLING VIA A CHANGE OF VARIABLES

STATISTICAL OPTIMIZATION OF COMPOSITE LAMINATES: INTRODUCTION OF COUPLING VIA A CHANGE OF VARIABLES STATISTICAL OPTIMIZATION OF COMPOSITE LAMINATES: INTRODUCTION OF COUPLING VIA A CHANGE OF VARIABLES Laurent Grosset,2 (AIAA member), Rodolphe Le Riche and Raphael T. Haftka 2 (AIAA fellow) CNRS URA 884

More information

Looking Under the EA Hood with Price s Equation

Looking Under the EA Hood with Price s Equation Looking Under the EA Hood with Price s Equation Jeffrey K. Bassett 1, Mitchell A. Potter 2, and Kenneth A. De Jong 1 1 George Mason University, Fairfax, VA 22030 {jbassett, kdejong}@cs.gmu.edu 2 Naval

More information

Black Box Search By Unbiased Variation

Black Box Search By Unbiased Variation Black Box Search By Unbiased Variation Per Kristian Lehre and Carsten Witt CERCIA, University of Birmingham, UK DTU Informatics, Copenhagen, Denmark ThRaSH - March 24th 2010 State of the Art in Runtime

More information

Finding Ground States of Sherrington-Kirkpatrick Spin Glasses with Hierarchical BOA and Genetic Algorithms

Finding Ground States of Sherrington-Kirkpatrick Spin Glasses with Hierarchical BOA and Genetic Algorithms Finding Ground States of Sherrington-Kirkpatrick Spin Glasses with Hierarchical BOA and Genetic Algorithms Martin Pelikan, Helmut G. Katzgraber and Sigismund Kobe MEDAL Report No. 284 January 28 Abstract

More information

CONTRASTING MAIN SELECTION METHODS IN GENETIC ALGORITHMS

CONTRASTING MAIN SELECTION METHODS IN GENETIC ALGORITHMS CONTRASTING MAIN SELECTION METHODS IN GENETIC ALGORITHMS Alfonso H., Cesan P., Fernandez N., Minetti G., Salto C., Velazco L. Proyecto UNLPAM-9/F9 Departamento de Informática - Facultad de Ingeniería Universidad

More information

Introduction. Genetic Algorithm Theory. Overview of tutorial. The Simple Genetic Algorithm. Jonathan E. Rowe

Introduction. Genetic Algorithm Theory. Overview of tutorial. The Simple Genetic Algorithm. Jonathan E. Rowe Introduction Genetic Algorithm Theory Jonathan E. Rowe University of Birmingham, UK GECCO 2012 The theory of genetic algorithms is beginning to come together into a coherent framework. However, there are

More information

Population-Based Incremental Learning with Immigrants Schemes in Changing Environments

Population-Based Incremental Learning with Immigrants Schemes in Changing Environments Population-Based Incremental Learning with Immigrants Schemes in Changing Environments Michalis Mavrovouniotis Centre for Computational Intelligence (CCI) School of Computer Science and Informatics De

More information

Efficient Discretization Scheduling in Multiple Dimensions. Laura A. Albert and David E. Goldberg. IlliGAL Report No February 2002

Efficient Discretization Scheduling in Multiple Dimensions. Laura A. Albert and David E. Goldberg. IlliGAL Report No February 2002 Efficient Discretization Scheduling in Multiple Dimensions Laura A. Albert and David E. Goldberg IlliGAL Report No. 2002006 February 2002 Illinois Genetic Algorithms Laboratory University of Illinois at

More information

1. Computação Evolutiva

1. Computação Evolutiva 1. Computação Evolutiva Renato Tinós Departamento de Computação e Matemática Fac. de Filosofia, Ciência e Letras de Ribeirão Preto Programa de Pós-Graduação Em Computação Aplicada 1.6. Aspectos Teóricos*

More information

Evolutionary computation

Evolutionary computation Evolutionary computation Andrea Roli andrea.roli@unibo.it DEIS Alma Mater Studiorum Università di Bologna Evolutionary computation p. 1 Evolutionary Computation Evolutionary computation p. 2 Evolutionary

More information

Evolutionary Computation. DEIS-Cesena Alma Mater Studiorum Università di Bologna Cesena (Italia)

Evolutionary Computation. DEIS-Cesena Alma Mater Studiorum Università di Bologna Cesena (Italia) Evolutionary Computation DEIS-Cesena Alma Mater Studiorum Università di Bologna Cesena (Italia) andrea.roli@unibo.it Evolutionary Computation Inspiring principle: theory of natural selection Species face

More information

Effects of the Use of Non-Geometric Binary Crossover on Evolutionary Multiobjective Optimization

Effects of the Use of Non-Geometric Binary Crossover on Evolutionary Multiobjective Optimization Effects of the Use of Non-Geometric Binary Crossover on Evolutionary Multiobjective Optimization Hisao Ishibuchi, Yusuke Nojima, Noritaka Tsukamoto, and Ken Ohara Graduate School of Engineering, Osaka

More information

Exploration of population fixed-points versus mutation rates for functions of unitation

Exploration of population fixed-points versus mutation rates for functions of unitation Exploration of population fixed-points versus mutation rates for functions of unitation J Neal Richter 1, Alden Wright 2, John Paxton 1 1 Computer Science Department, Montana State University, 357 EPS,

More information

Stochastic Local Search in Continuous Domains

Stochastic Local Search in Continuous Domains Stochastic Local Search in Continuous Domains Questions to be Answered When Designing A Novel Algorithm Petr Pošík Czech Technical University in Prague Faculty of Electrical Engineering, Department of

More information

Fitness distributions and GA hardness

Fitness distributions and GA hardness Fitness distributions and GA hardness Yossi Borenstein and Riccardo Poli Department of Computer Science University of Essex Abstract. Considerable research effort has been spent in trying to formulate

More information

Binary Particle Swarm Optimization with Crossover Operation for Discrete Optimization

Binary Particle Swarm Optimization with Crossover Operation for Discrete Optimization Binary Particle Swarm Optimization with Crossover Operation for Discrete Optimization Deepak Singh Raipur Institute of Technology Raipur, India Vikas Singh ABV- Indian Institute of Information Technology

More information

A Non-Parametric Statistical Dominance Operator for Noisy Multiobjective Optimization

A Non-Parametric Statistical Dominance Operator for Noisy Multiobjective Optimization A Non-Parametric Statistical Dominance Operator for Noisy Multiobjective Optimization Dung H. Phan and Junichi Suzuki Deptartment of Computer Science University of Massachusetts, Boston, USA {phdung, jxs}@cs.umb.edu

More information

Lecture 22. Introduction to Genetic Algorithms

Lecture 22. Introduction to Genetic Algorithms Lecture 22 Introduction to Genetic Algorithms Thursday 14 November 2002 William H. Hsu, KSU http://www.kddresearch.org http://www.cis.ksu.edu/~bhsu Readings: Sections 9.1-9.4, Mitchell Chapter 1, Sections

More information

Chapter 8: Introduction to Evolutionary Computation

Chapter 8: Introduction to Evolutionary Computation Computational Intelligence: Second Edition Contents Some Theories about Evolution Evolution is an optimization process: the aim is to improve the ability of an organism to survive in dynamically changing

More information

Grammar Transformations in an EDA for Genetic Programming

Grammar Transformations in an EDA for Genetic Programming Grammar Transformations in an EDA for Genetic Programming Peter A.N. Bosman Edwin D. de Jong institute of information and computing sciences, utrecht university technical report UU-CS-2004-047 www.cs.uu.nl

More information

Evolutionary computation

Evolutionary computation Evolutionary computation Andrea Roli andrea.roli@unibo.it Dept. of Computer Science and Engineering (DISI) Campus of Cesena Alma Mater Studiorum Università di Bologna Outline 1 Basic principles 2 Genetic

More information

Why Topology Matters. Spatial Evolutionary Algorithms Evolution in Space and Time. Island Population Topologies. Main Population Topologies

Why Topology Matters. Spatial Evolutionary Algorithms Evolution in Space and Time. Island Population Topologies. Main Population Topologies Why Topology Matters Spatial Evolutionary Algorithms Evolution in Space and Time Marco Tomassini marco.tomassini@unil.ch University of Lausanne, Switzerland The spatial structure of a population will be

More information

An Evolution Strategy for the Induction of Fuzzy Finite-state Automata

An Evolution Strategy for the Induction of Fuzzy Finite-state Automata Journal of Mathematics and Statistics 2 (2): 386-390, 2006 ISSN 1549-3644 Science Publications, 2006 An Evolution Strategy for the Induction of Fuzzy Finite-state Automata 1,2 Mozhiwen and 1 Wanmin 1 College

More information

The Role of Crossover in Genetic Algorithms to Solve Optimization of a Function Problem Falih Hassan

The Role of Crossover in Genetic Algorithms to Solve Optimization of a Function Problem Falih Hassan The Role of Crossover in Genetic Algorithms to Solve Optimization of a Function Problem Falih Hassan ABSTRACT The genetic algorithm is an adaptive search method that has the ability for a smart search

More information

Mid-term Exam Answers and Final Exam Study Guide CIS 675 Summer 2010

Mid-term Exam Answers and Final Exam Study Guide CIS 675 Summer 2010 Mid-term Exam Answers and Final Exam Study Guide CIS 675 Summer 2010 Midterm Problem 1: Recall that for two functions g : N N + and h : N N +, h = Θ(g) iff for some positive integer N and positive real

More information

Rule Acquisition for Cognitive Agents by Using Estimation of Distribution Algorithms

Rule Acquisition for Cognitive Agents by Using Estimation of Distribution Algorithms Rule cquisition for Cognitive gents by Using Estimation of Distribution lgorithms Tokue Nishimura and Hisashi Handa Graduate School of Natural Science and Technology, Okayama University Okayama 700-8530,

More information

Probability (Phi) Proportion BBs (Pbb) 1. Proportion BBs (Pbb) Pop Size (n) Pop Size (n)

Probability (Phi) Proportion BBs (Pbb) 1. Proportion BBs (Pbb) Pop Size (n) Pop Size (n) THE POPULATION SIZING PROBLEM IN (P)GAS Zden»ek Konfr»st Department of Cybernetics, Czech Technical University Abstract One of interesting problems in genetic algorithms (GAs) is the population sizing

More information

Finding an Impulse Response Function Using Genetic Programming

Finding an Impulse Response Function Using Genetic Programming Finding an Impulse Response Function Using Genetic Programming Martin A. Keane John R. Koza James P. Rice Third Millennium Venture Capital Limited 5733 Grover Chicago, Illinois 663 32-777524 Computer Science

More information

Unsupervised Fuzzy Tournament Selection

Unsupervised Fuzzy Tournament Selection Applied Mathematical Sciences, Vol. 5, 2011, no. 58, 2863-2881 Unsupervised Fuzzy Tournament Selection Khalid Jebari, Abdelaziz El moujahid, Amina Dik, Abdelaziz Bouroumi and Aziz Ettouhami Laboratoire

More information

Quad-trees: A Data Structure for Storing Pareto-sets in Multi-objective Evolutionary Algorithms with Elitism

Quad-trees: A Data Structure for Storing Pareto-sets in Multi-objective Evolutionary Algorithms with Elitism Quad-trees: A Data Structure for Storing Pareto-sets in Multi-objective Evolutionary Algorithms with Elitism Sanaz Mostaghim 1 and Jürgen Teich 2 1 Electrical Engineering Department University of Paderborn,

More information

TAKEOVER TIME IN PARALLEL POPULATIONS WITH MIGRATION

TAKEOVER TIME IN PARALLEL POPULATIONS WITH MIGRATION TAKEOVER TIME IN PARALLEL POPULATIONS WITH MIGRATION Günter Rudolph Universität Dortmund Fachbereich Informatik Lehrstuhl für Algorithm Engineering 441 Dortmund / Germany Guenter.Rudolph@uni-dortmund.de

More information

Genetic Algorithm: introduction

Genetic Algorithm: introduction 1 Genetic Algorithm: introduction 2 The Metaphor EVOLUTION Individual Fitness Environment PROBLEM SOLVING Candidate Solution Quality Problem 3 The Ingredients t reproduction t + 1 selection mutation recombination

More information

Plateaus Can Be Harder in Multi-Objective Optimization

Plateaus Can Be Harder in Multi-Objective Optimization Plateaus Can Be Harder in Multi-Objective Optimization Tobias Friedrich and Nils Hebbinghaus and Frank Neumann Max-Planck-Institut für Informatik, Campus E1 4, 66123 Saarbrücken, Germany Abstract In recent

More information

Evolving more efficient digital circuits by allowing circuit layout evolution and multi-objective fitness

Evolving more efficient digital circuits by allowing circuit layout evolution and multi-objective fitness Evolving more efficient digital circuits by allowing circuit layout evolution and multi-objective fitness Tatiana Kalganova Julian Miller School of Computing School of Computing Napier University Napier

More information

Forecasting & Futurism

Forecasting & Futurism Article from: Forecasting & Futurism December 2013 Issue 8 A NEAT Approach to Neural Network Structure By Jeff Heaton Jeff Heaton Neural networks are a mainstay of artificial intelligence. These machine-learning

More information

Evolution Strategies for Constants Optimization in Genetic Programming

Evolution Strategies for Constants Optimization in Genetic Programming Evolution Strategies for Constants Optimization in Genetic Programming César L. Alonso Centro de Inteligencia Artificial Universidad de Oviedo Campus de Viesques 33271 Gijón calonso@uniovi.es José Luis

More information

Cross Entropy and Adaptive Variance Scaling in Continuous EDA

Cross Entropy and Adaptive Variance Scaling in Continuous EDA Cross Entropy and Adaptive Variance Scaling in Continuous EDA Cai Yunpeng caiyp78@gmail.com Xu Hua xuhua@tsinghua.edu.cn Sun Xiaomin sxm23@tsinghua.edu.cn Jia Peifa dcsjpf@tsinghua.edu.cn State Key Lab

More information

Epistasis in Multi-Objective Evolutionary Recurrent Neuro- Controllers

Epistasis in Multi-Objective Evolutionary Recurrent Neuro- Controllers Brock University Department of Computer Science Epistasis in Multi-Objective Evolutionary Recurrent Neuro- Controllers Mario Ventresca and Beatrice Ombuki-Berman Technical Report # CS-06-06 November 2006

More information

Computational statistics

Computational statistics Computational statistics Combinatorial optimization Thierry Denœux February 2017 Thierry Denœux Computational statistics February 2017 1 / 37 Combinatorial optimization Assume we seek the maximum of f

More information

Different Criteria for Active Learning in Neural Networks: A Comparative Study

Different Criteria for Active Learning in Neural Networks: A Comparative Study Different Criteria for Active Learning in Neural Networks: A Comparative Study Jan Poland and Andreas Zell University of Tübingen, WSI - RA Sand 1, 72076 Tübingen, Germany Abstract. The field of active

More information

Automatic discovery of ranking formulas for playing with multi-armed bandits

Automatic discovery of ranking formulas for playing with multi-armed bandits Automatic discovery of ranking formulas for playing with multi-armed bandits Francis Maes, Louis Wehenkel, and Damien Ernst University of Liège Dept. of Electrical Engineering and Computer Science Institut

More information

Algorithmic probability, Part 1 of n. A presentation to the Maths Study Group at London South Bank University 09/09/2015

Algorithmic probability, Part 1 of n. A presentation to the Maths Study Group at London South Bank University 09/09/2015 Algorithmic probability, Part 1 of n A presentation to the Maths Study Group at London South Bank University 09/09/2015 Motivation Effective clustering the partitioning of a collection of objects such

More information

Applying Price s Equation to Survival Selection

Applying Price s Equation to Survival Selection Applying Price s Equation to Survival Selection Jeffrey K. Bassett George Mason University Krasnow Institute, MS2A1 4400 University Dr. Fairfax, VA 22030 jbassett@cs.gmu.edu Mitchell A. Potter Naval Research

More information

Compact representations as a search strategy: Compression EDAs

Compact representations as a search strategy: Compression EDAs Compact representations as a search strategy: Compression EDAs Marc Toussaint Institute for Adaptive and Neural Computation, University of Edinburgh, 5 Forrest Hill, Edinburgh EH1 2QL, Scotland, UK Abstract

More information

Sorting Network Development Using Cellular Automata

Sorting Network Development Using Cellular Automata Sorting Network Development Using Cellular Automata Michal Bidlo, Zdenek Vasicek, and Karel Slany Brno University of Technology, Faculty of Information Technology Božetěchova 2, 61266 Brno, Czech republic

More information

optimization problem x opt = argmax f(x). Definition: Let p(x;t) denote the probability of x in the P population at generation t. Then p i (x i ;t) =

optimization problem x opt = argmax f(x). Definition: Let p(x;t) denote the probability of x in the P population at generation t. Then p i (x i ;t) = Wright's Equation and Evolutionary Computation H. Mühlenbein Th. Mahnig RWCP 1 Theoretical Foundation GMD 2 Laboratory D-53754 Sankt Augustin Germany muehlenbein@gmd.de http://ais.gmd.de/οmuehlen/ Abstract:

More information

On the errors introduced by the naive Bayes independence assumption

On the errors introduced by the naive Bayes independence assumption On the errors introduced by the naive Bayes independence assumption Author Matthijs de Wachter 3671100 Utrecht University Master Thesis Artificial Intelligence Supervisor Dr. Silja Renooij Department of

More information

Genetic Algorithm for Solving the Economic Load Dispatch

Genetic Algorithm for Solving the Economic Load Dispatch International Journal of Electronic and Electrical Engineering. ISSN 0974-2174, Volume 7, Number 5 (2014), pp. 523-528 International Research Publication House http://www.irphouse.com Genetic Algorithm

More information

Robust Multi-Objective Optimization in High Dimensional Spaces

Robust Multi-Objective Optimization in High Dimensional Spaces Robust Multi-Objective Optimization in High Dimensional Spaces André Sülflow, Nicole Drechsler, and Rolf Drechsler Institute of Computer Science University of Bremen 28359 Bremen, Germany {suelflow,nd,drechsle}@informatik.uni-bremen.de

More information

Metaheuristics and Local Search

Metaheuristics and Local Search Metaheuristics and Local Search 8000 Discrete optimization problems Variables x 1,..., x n. Variable domains D 1,..., D n, with D j Z. Constraints C 1,..., C m, with C i D 1 D n. Objective function f :

More information

When to Use Bit-Wise Neutrality

When to Use Bit-Wise Neutrality When to Use Bit-Wise Neutrality Tobias Friedrich Department 1: Algorithms and Complexity Max-Planck-Institut für Informatik Saarbrücken, Germany Frank Neumann Department 1: Algorithms and Complexity Max-Planck-Institut

More information

Biology as Information Dynamics

Biology as Information Dynamics Biology as Information Dynamics John Baez Stanford Complexity Group April 20, 2017 What is life? Self-replicating information! Information about what? How to self-replicate! It is clear that biology has

More information

Combinatorial Optimization by Learning and Simulation of Bayesian Networks

Combinatorial Optimization by Learning and Simulation of Bayesian Networks UNCERTAINTY IN ARTIFICIAL INTELLIGENCE PROCEEDINGS 2000 343 Combinatorial Optimization by Learning and Simulation of Bayesian Networks Pedro Larraiiaga, Ramon Etxeberria, Jose A. Lozano and Jose M. Peiia

More information

A Review of Stochastic Local Search Techniques for Theorem Provers

A Review of Stochastic Local Search Techniques for Theorem Provers A Review of Stochastic Local Search Techniques for Theorem Provers Termpaper presentation for CSE814-Formal Methods in Software Development Course Michigan State University, East Lansing, USA Presenter:

More information

The Evolution of Sex Chromosomes through the. Baldwin Effect

The Evolution of Sex Chromosomes through the. Baldwin Effect The Evolution of Sex Chromosomes through the Baldwin Effect Larry Bull Computer Science Research Centre Department of Computer Science & Creative Technologies University of the West of England, Bristol

More information

Crossover Gene Selection by Spatial Location

Crossover Gene Selection by Spatial Location Crossover Gene Selection by Spatial Location ABSTRACT Dr. David M. Cherba Computer Science Department Michigan State University 3105 Engineering Building East Lansing, MI 48823 USA cherbada@cse.msu.edu

More information