MICROCANONICAL OPTIMIZATION APPLIED TO THE TRAVELING SALESMAN PROBLEM

Similar documents
Single Solution-based Metaheuristics

5. Simulated Annealing 5.1 Basic Concepts. Fall 2010 Instructor: Dr. Masoud Yaghini

SIMU L TED ATED ANNEA L NG ING

Motivation, Basic Concepts, Basic Methods, Travelling Salesperson Problem (TSP), Algorithms

Design and Analysis of Algorithms

Artificial Intelligence Heuristic Search Methods

Optimization Methods via Simulation

PROBLEM SOLVING AND SEARCH IN ARTIFICIAL INTELLIGENCE

Hill climbing: Simulated annealing and Tabu search

Methods for finding optimal configurations

( ) ( ) ( ) ( ) Simulated Annealing. Introduction. Pseudotemperature, Free Energy and Entropy. A Short Detour into Statistical Mechanics.

Introduction to Simulated Annealing 22c:145

Algorithm Design Strategies V

7.1 Basis for Boltzmann machine. 7. Boltzmann machines

Simulated Annealing. Local Search. Cost function. Solution space

A Two-Stage Simulated Annealing Methodology

Algorithms and Complexity theory

Zebo Peng Embedded Systems Laboratory IDA, Linköping University

Local Search & Optimization

Simulated Annealing applied to the Traveling Salesman Problem. Eric Miller

Heuristic Optimisation

Lin-Kernighan Heuristic. Simulated Annealing

Methods for finding optimal configurations

Lecture H2. Heuristic Methods: Iterated Local Search, Simulated Annealing and Tabu Search. Saeed Bastani

A.I.: Beyond Classical Search

Markov Chain Monte Carlo. Simulated Annealing.

1 Heuristics for the Traveling Salesman Problem

6. APPLICATION TO THE TRAVELING SALESMAN PROBLEM

Simulated Annealing. 2.1 Introduction

1a. Introduction COMP6741: Parameterized and Exact Computation

Ant Colony Optimization: an introduction. Daniel Chivilikhin

CS 331: Artificial Intelligence Local Search 1. Tough real-world problems

Local Search & Optimization

Overview. Optimization. Easy optimization problems. Monte Carlo for Optimization. 1. Survey MC ideas for optimization: (a) Multistart

SYSTEMS SCIENCE AND CYBERNETICS Vol. III Simulated Annealing: From Statistical Thermodynamics to Combinatory Problems Solving - D.

Simulated Annealing for Constrained Global Optimization

Quantum annealing for problems with ground-state degeneracy

Metaheuristics and Local Search

Local and Stochastic Search

Acceptance Driven Local Search and Evolutionary Algorithms

Fundamentals of Metaheuristics

Metaheuristics and Local Search. Discrete optimization problems. Solution approaches

Solving the Homogeneous Probabilistic Traveling Salesman Problem by the ACO Metaheuristic

Finding optimal configurations ( combinatorial optimization)

12. LOCAL SEARCH. gradient descent Metropolis algorithm Hopfield neural networks maximum cut Nash equilibria

Spin Glas Dynamics and Stochastic Optimization Schemes. Karl Heinz Hoffmann TU Chemnitz

Metaheuristics. 2.3 Local Search 2.4 Simulated annealing. Adrian Horga

3D HP Protein Folding Problem using Ant Algorithm

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling

Unit 1A: Computational Complexity

Thermodynamical Approach to the Traveling Salesman Problem: An Efficient Simulation Algorithm I

CS 781 Lecture 9 March 10, 2011 Topics: Local Search and Optimization Metropolis Algorithm Greedy Optimization Hopfield Networks Max Cut Problem Nash

Generalization of Dominance Relation-Based Replacement Rules for Memetic EMO Algorithms

Scaling Up. So far, we have considered methods that systematically explore the full search space, possibly using principled pruning (A* etc.).

Modern WalkSAT algorithms

Parameter estimation using simulated annealing for S- system models of biochemical networks. Orland Gonzalez

The Traveling Salesman Problem: A Neural Network Perspective. Jean-Yves Potvin

Informatik-Bericht Nr

CS264: Beyond Worst-Case Analysis Lecture #4: Parameterized Analysis of Online Paging

Efficient Cryptanalysis of Homophonic Substitution Ciphers

12. LOCAL SEARCH. gradient descent Metropolis algorithm Hopfield neural networks maximum cut Nash equilibria

Microcanonical Mean Field Annealing: A New Algorithm for Increasing the Convergence Speed of Mean Field Annealing.

Heuristics for The Whitehead Minimization Problem

Minicourse on: Markov Chain Monte Carlo: Simulation Techniques in Statistics

Introduction. An Introduction to Algorithms and Data Structures

Sensitive Ant Model for Combinatorial Optimization

Computational statistics

MONTE CARLO METHODS IN SEQUENTIAL AND PARALLEL COMPUTING OF 2D AND 3D ISING MODEL

arxiv: v1 [cond-mat.stat-mech] 5 Jun 2008

CMOS Ising Computer to Help Optimize Social Infrastructure Systems

Tabu Search. Biological inspiration is memory the ability to use past experiences to improve current decision making.

On-Line Load Balancing

Hertz, Krogh, Palmer: Introduction to the Theory of Neural Computation. Addison-Wesley Publishing Company (1991). (v ji (1 x i ) + (1 v ji )x i )

An Effective Chromosome Representation for Evolving Flexible Job Shop Schedules

Artificial Intelligence. 3 Problem Complexity. Prof. Dr. Jana Koehler Fall 2016 HSLU - JK

Statistics and Quantum Computing

arxiv: v3 [physics.comp-ph] 22 Sep 2016

Gaussian Distributions and Global. Optimization: An exploration into improved. performance

Module 1: Analyzing the Efficiency of Algorithms

Data Structures in Java

Week Cuts, Branch & Bound, and Lagrangean Relaxation

MCMC Simulated Annealing Exercises.

A Survey on Travelling Salesman Problem

Algorithms. NP -Complete Problems. Dong Kyue Kim Hanyang University

Doubly-Rooted Stem-and-Cycle Ejection Chain Algorithm for the Asymmetric Traveling Salesman Problem

Random Walks A&T and F&S 3.1.2

Advanced sampling. fluids of strongly orientation-dependent interactions (e.g., dipoles, hydrogen bonds)

A Hybrid Simulated Annealing with Kempe Chain Neighborhood for the University Timetabling Problem

A Self-Stabilizing Algorithm for Finding a Minimal Distance-2 Dominating Set in Distributed Systems

Static Load-Balancing Techniques for Iterative Computations on Heterogeneous Clusters

MatSci 331 Homework 4 Molecular Dynamics and Monte Carlo: Stress, heat capacity, quantum nuclear effects, and simulated annealing

A pruning pattern list approach to the permutation flowshop scheduling problem

Comments on An Improvement to the Brent s Method

Intuitionistic Fuzzy Estimation of the Ant Methodology

4 th ECADA Evolutionary Computation for the Automated Design of Algorithms GECCO WORKSHOP 2014

Integer weight training by differential evolution algorithms

Optimisation and Operations Research

arxiv:cond-mat/ v1 [cond-mat.mtrl-sci] 28 Oct 2004

Physics 403. Segev BenZvi. Numerical Methods, Maximum Likelihood, and Least Squares. Department of Physics and Astronomy University of Rochester

Artificial Intelligence Methods (G5BAIM) - Examination

Transcription:

International Journal of Modern Physics C, Vol. 9, No. 1 (1998) 133 146 c World Scientific Publishing Company MICROCANONICAL OPTIMIZATION APPLIED TO THE TRAVELING SALESMAN PROBLEM ALEXANDRE LINHARES Computação Aplicada e Automação, UFF 2421-24 Niterói, RJ, Brazil E-mail: linhares@nucleo.inpe.br JOSÉ R.A.TORREÃO Computação Aplicada e Automação, UFF 2421-24 Niterói, RJ, Brazil E-mail: jrat@caa.uff.br Received 24 October 1997 Revised 17 December 1997 Optimization strategies based on simulated annealing and its variants have been extensively applied to the traveling salesman problem (TSP). Recently, there has appeared a new physics-based metaheuristic, called the microcanonical optimization algorithm (µo), which does not resort to annealing, and which has proven a superior alternative to the annealing procedures in various applications. Here we present the first performance evaluation of µo as applied to the TSP. When compared to three annealing strategies (simulated annealing, microcanonical annealing and Tsallis annealing), and to a tabu search algorithm, the microcanonical optimization has yielded the best overall results for several instances of the euclidean TSP. This confirms µo as a competitive approach for the solution of general combinatorial optimization problems. Keywords: Combinatorial Optimization; Microcanonical Ensemble; Simulated Annealing; Traveling Salesman Problem. 1. Introduction The traveling salesman problem (TSP) has been studied since the early days of scientific computation, and is now considered the benchmark in the field of combinatorial optimization. The problem can be easily stated: given a set of cities, the goal is to find a path of minimal cost, going through each city only once and returning to the starting point. In spite of its simple formulation, the TSP has been proven to be NP-Hard, meaning that there probably does not exist an algorithm which can exactly solve a general instance of the problem in plausible processing time. The best that can be expected is thus to find approximate strategies of solution, called heuristics. If a heuristic is a general-purpose procedure which can be applied to a variety of problems, it is referred to as a metaheuristic. 133

134 A. Linhares & J. R. A. Torreão Among the metaheuristics employed for the TSP, optimization algorithms derived from statistical physics have received a great deal of attention. 1 3 Simulated annealing, as introduced by Kirkpatrick et al., 4 was the first such algorithm, and many variants of it have appeared, such as fast simulated annealing, microcanonical annealing, 6 and Tsallis annealing. 3 Recently, a new strategy has been proposed which is also based on principles of statistical physics, but which does not resort to annealing. It is called the microcanonical optimization algorithm (µo), and has so far been employed, with remarkable success, in the context of visual processing, 7,8 and for task allocation in distributed systems. 9 Here, we present an analysis of µo when applied to the TSP, comparing it to some annealing-based procedures (simulated annealing, microcanonical annealing and Tsallis annealing), and also to a tabu search algorithm. 1 The results which we report show µo to be a very competitive metaheuristic in this domain: when considering both execution time and solution quality, it yielded the best performance of all the evaluated algorithms. In the following section, we describe the microcanonical optimization algorithm. Next, we discuss some implementation details of the alternative metaheuristics considered. In Sec. 4, we present and analyze the results obtained in our work, concluding with our final remarks in Sec.. 2. Microcanonical Optimization The microcanonical optimization algorithm consists of two procedures which are alternately applied: initialization and sampling. The initialization implements a local and optionally aggressive search of the solution space, in order to reach a local-minimum configuration. From there, the sampling phase proceeds, trying to free the solution from the local minimum, by taking it to another configuration of equivalent cost. One can picture the metaheuristic, once stuck in a local-minimum valley, as trying to evolve by going around the peaks in the solution space, instead of attempting to climb them, as in simulated annealing, for instance. This is done by resorting to the microcanonical simulation algorithm by Creutz, 11 which generates samples of fixed-energy configurations (see below). After the sampling phase, a new initialization is run and the algorithm thus proceeds, alternating between the two phases, until a stopping condition is reached. In what follows, we treat in greater detail the two phases of the microcanonical optimization. A pseudocode for the algorithm is given in Appendix A. 2.1. Initialization In the initialization, µo performs a local search, starting from an arbitrary solution and proposing moves which are accepted only when leading to configurations of lower cost (lower energy, in physical terms). Optionally, an aggressive implementation of this phase can be chosen, meaning that the algorithm will always pick the best candidate in a subset of possible moves.

Microcanonical Optimization Applied to the Traveling Salesman Problem 13 In a non-aggressive implementation, the only free parameter of the initialization phase defines its stopping condition: since it cannot be rigorously established when a local minimum has been reached, it is necessary to define a maximum number of rejected moves as the criterium for interrupting this phase. In the case of an aggressive implementation (which we chose), it is also necessary to define the number of candidate moves to be considered in each initialization step (, in our work). We also remark that, for the definition of the parameters to be employed in the sampling phase (see below), a list may be compiled, in the initialization, of those moves which have been rejected for leading to higher costs when compared to the current solution. 2.2. Sampling As already mentioned, in the sampling phase the µo metaheuristic tries to free itself from the local minimum reached in the initialization, at the same time trying to remain close, in terms of cost, to the best solution so far obtained. It implements, for this purpose, a version of the Creutz algorithm, assuming an extra degree of freedom, called the demon, which generates small perturbations on the current solution. At each sampling iteration, random moves are proposed which are accepted only if the demon is capable of yielding or absorbing the cost difference incurred. In µo, the demon is defined by two parameters: its capacity, D MAX,andits initial value, D I. The sampling generates a sequence of states whose energy is conserved, except for small fluctuations which are modeled by the demon. Calling E S the energy (cost) of the solution obtained in the initialization, and D and E the energy of the demon and of the solution, respectively, at a given instant in the sampling phase, we must have E + D = E S + D I = constant. Thus, in terms of the initial energy and the capacity of the demon, this phase generates solutions in the cost interval [E S D MAX + D I,E S +D I ]. D I and D MAX are, therefore, the main parameters to be considered in the implementation of the sampling. In the original formulation of the algorithm, such parameters were taken, at each sampling phase, as fixed fractions of the final cost obtained in the previous initialization. 7 As one of the contributions of the present work, we have proposed an adaptive strategy for the determination of such parameters: taking the list of rejected movements compiled in the initialization phase (see above), we have sorted it in growing order of the cost jumps, choosing two of its lower entries as the values of demon capacity and initial energy. The idea is that such values will be representative of the hills found in the landscape of the region being searched in the solution space, and will thus be adequate for defining the magnitude of the perturbations required for the evolution of the current solution, in the sampling phase. In our implementations of µo for the TSP, the initialization was executed until a count of 1n consecutively rejected moves was reached, where n was the number

136 A. Linhares & J. R. A. Torreão of cities in the problem. The values of D MAX and D I were both usually taken as equal to the th lowest entry in the list of rejected moves compiled in the initialization, except for a certain kind of city distribution which required a change in this prescription (see Sec. 4). The sampling phase was run for only iterations, and the algorithm was made to stop when reaching a count of 1 moves without improvement in the best solution encountered. 3. Alternative Strategies In our experiments, we compared the performance of µo to those yielded by alternative strategies: simulated annealing, microcanonical annealing, Tsallis annealing and tabu search. Here we discuss some of the features of the implementation of such algorithms in our work. 3.1. Simulated annealing (SA) Simulated annealing, as proposed by Kirkpatrick et al., 4 consists in the iterated implementation of the Metropolis algorithm, 12 for a sequence of decreasing temperatures. The Metropolis algorithm is a computational procedure, long known in statistical physics, which generates samples of the states of a physical system at a fixed temperature. Since such a system obeys the Gibbs distribution, the states generated at low temperatures will be low energy states. 13,14 Identifying the energy of the system with the cost function in an optimization problem, Kirkpatrick et al. proposed the following optimization strategy: Starting from an arbitrary solution, and a high temperature, the Metropolis algorithm is implemented, which means that moves are proposed which are accepted with probability p =min(1,exp ( E/T)), where E is the cost variation incurred, and T is the current temperature. After a large number of iterations, the value of T is decreased, and the process is repeated until T. The initial value and rate of decrease of the temperature (which has no physical meaning in the optimization, being just a global control parameter of the process) constitute the annealing schedule of the algorithm. In our implementations, we followed the prescriptions by Cerny, 1 taking the temperature to decrease by 7% of its value at each annealing step, and keeping it constant for 1n accepted moves or 1n rejected moves, whichever came first, with n being the number of cities in the problem. The initial temperature was empirically determined: 1 trial moves, starting from the initial random solution, were analyzed, and the initial temperature was chosen greater than the maximum cost variation observed. 3.2. Tsallis annealing This corresponds to a variant of simulated annealing, based on the statistics proposed by C. Tsallis. 16 Here, the acceptance probability of the Metropolis algorithm is generalized to p =min(1,[1 (1 q) E/T] 1/(1 q) ), such that SA is recovered in the limit of q 1. By appropriately choosing the value of q, it has been claimed 3

Microcanonical Optimization Applied to the Traveling Salesman Problem 137 that this algorithm can produce plausible TSP solutions in fewer steps than with fast simulated annealing. In our implementations, we followed the general annealing prescriptions described above for SA. As for the parameter q, specific to the Tsallis annealing, it has been suggested that the algorithm improves, in what concerns execution times, as q decreases towards. 3 Such general behavior was confirmed in our work, but, even though an exhaustive analysis has not been undertaken, we noticed a corresponding degradation in solution quality for q< 1. The value q = 1 was therefore employed in our experiments. 3.3. Microcanonical annealing (MA) This algorithm also corresponds to a variant of simulated annealing, now based on a simulation of the states of a physical system at fixed energy, through the Creutz algorithm, 6 instead of at fixed temperature (The SA and Tsallis algorithms would thus correspond to canonical annealings). As originally proposed for visual processing applications, MA employed a lattice of demons, and was suited only for parallel implementations. In our single-demon sequential version, microcanonical annealing consists, basically, in the iterative application of the Creutz algorithm for progressively lower values of demon capacity. In our implementations, we took a demon of zero initial energy, such that, at the ith annealing step, states would be generated in the cost interval [E (i 1) D (i),e (i 1) ], where D (i) represents the current demon capacity, and E (i 1) represents the final energy reached in the previous annealing step. The rate of decrease of the demon capacity was the same used in the canonical annealings for temperature decrease, with the initial demon value determined similarly to the initial annealing temperature: starting from a random solution, 1 prospective moves were analyzed, and the largest cost variation was taken as the demon capacity in the first annealing step. 3.4. Tabu search In order to avoid getting entrapped in a local minimum, the tabu search algorithm selects, at each step, the best of a certain number of candidate moves (, in our implementations), even if it leads to a higher cost, in which case the corresponding reverse move is included in a tabu list, to prevent the return to a solution already considered. In our experiments, we worked with a tabu list of 7 moves, following the suggestion of Glover, 1 with each new tabu move being included in a random position in the list, so that its interdiction period would also be random. Another feature of our implementations was a so-called aspiration criterium, according to which, if a given tabu move leads to a solution which tops the best one so far encountered, its interdiction is ignored. The tabu search was made to stop at a count of moves without improvement.

138 A. Linhares & J. R. A. Torreão 4. Experiments Our performance evaluation of µo was based on the solution of several instances of the euclidean TSP, employing a path-reversal dynamics. 17 This means that the solution cost was taken as the total tour length measured by the euclidean norm, and that each trial move was a replacement of a randomly selected section of the tour by its reverse. Results obtained with a Pentium 133 processor will be reported here, for the following city distributions: P1: 1 cities organized in a rectangular grid. Such a distribution, also employed by Cerny 1 and by Laarhoven and Aarts, 14 displays a global minimum which can be easily perceived, and is an example of degenerate topology, allowing many solutions of the same cost. P3: 3 cities randomly distributed in eight distinct clusters along the sides of a square region. The optimal path which is not known a priori must cross each cluster only once. PR76, PR124 and PR423: Configurations of 76, 124 and 423 cities, respectively, proposed by Padberg and Rinaldi, and compiled in the TSPLIB library. 18 The corresponding optimal solutions are also shown in the TSPLIB. K: Configuration of cities proposed by Krolak and also found, along with its optimal solution, in the TSPLIB. In order to appreciate the quality of the solutions yielded by the various algorithms, we considered the distribution of the results obtained in several runs. The frequency histograms of the final costs for P1 and P3, in executions, are shown in Figs. 1 and 2, where we include the results for the iterative improvement algorithm, which corresponds to implementing only the non-aggressive initialization phase of µo. From the figures, the superiority of the microcanonical optimization over the other approaches is apparent, but the tabu search and microcanonical annealing methods also proved to be competitive. SA and Tsallis annealing yielded poorer quality solutions, even though the latter was very fast. Table 1 gives an idea of the average running times involved. It is important to remark that, due to the peculiarities of implementation of each algorithm, some of them tend naturally to prolong their execution in comparison to others. For instance, µo and tabu search will only stop after reaching a certain number of iterations without improvement, which means that, even after a long period without any progress, once those algorithms find a better configuration, they are granted an additional running time (of 1 iterations for µo, and for tabu search). The same is not true of the annealing strategies, which have their running times linked to fixed annealing schedules.

Microcanonical Optimization Applied to the Traveling Salesman Problem 139 Iterative Improvement Simulated Annealing 8 8 6 4 6 4 37 3823 394 468 37 3823 394 468 Microcanonical Annealing Tsallis Annealing 8 8 6 4 37 3823 394 468 6 4 37 3823 394 468 Tabu Search Microcanonical Optimization 8 8 6 4 6 4 37 3823 394 468 37 3823 394 468 Fig. 1., in fifty runs, of the final costs obtained for Problem P1. 3 1 3 1 Microcanonical Annealing 27 31 347 38 3 1 Iterative Improvement 27 31 347 38 Tabu Search 27 31 347 38 3 1 3 1 3 1 Simulated Annealing 27 31 347 38 Tsallis Annealing 27 31 347 38 Microcanonical Optimization 27 31 347 38 Fig. 2., in fifty runs, of the final costs obtained for Problem P3.

14 A. Linhares & J. R. A. Torreão Table 1. Average execution time (minutes), in five runs of µo, tabu search, microcanonical annealing (MA), Tsallis annealing, and simulated annealing (SA). Processor: Pentium 133. µ Tabu MA Tsallis SA P1 :48 :8 :47 1: 2:37 P3 2: 3:29 2:9 1:48 4:4 From such initial results, we have been led to undertake a more careful comparative analysis of µo, tabu search and microcanonical annealing. Table 2 summarizes the results obtained in runs for the distributions K and PR76. The corresponding graphs of running time versus final cost for K are depicted in Fig. 3. We see that the microcanonical annealing did not show any appreciable variation in execution time, even though it performed quite poorly, in this respect, in problem K. Tabu search, on the other hand, showed a behavior similar to that of µo, a feature which was observed for all configurations where the cities were evenly distributed over the plane, without the formation of well-defined groups. The solutions yielded by µo were slightly superior to those generated by the annealing, but required a little longer processing time in K. A different situation was met in problems PR124 and PR439, which share the peculiar characteristic of presenting relatively distant groups of densely packed cities, in a topology quite distinct from the ones previously analyzed. Such topology gives rise to the existence of a large number of local-minimum solutions, differing only in the intra-group sequences of cities, which are very close in cost. In this kind of problem, the intrinsic divide-and-conquer nature of annealing 4,6 proves to be quite invaluable, since it allows the initial optimization of the long paths between groups which are dominant in terms of cost leaving the finer details of Table 2. Average, maximum, and minimum values obtained in runs of µo, tabu search, and microcanonical annealing (MA), for problems K and PR76. E means cost and t means execution time, in minutes. Processor: Pentium 133. K E avg E min E max t avg t min t max µo 316 29696 3941 2: :9 4: Tabu 3392 29869 311 1:28 :2 2:3 MA 3271 29771 319 8:42 8:33 8:48 PR76 E avg E min E max t avg t min t max µo 1837 189 198 3:42 2:34 :43 Tabu 18736 189 19921 :38 2:37 11:43 MA 19418 189 1111 3:32 3:27 3:37

Microcanonical Optimization Applied to the Traveling Salesman Problem 141 :11:31 Microcanonical Optimization :1: :8:38 :7:12 Time ::46 :4:19 :2:3 :1:26 :: 294 299 34 39 314 319 324 329 334 339 344 :11:31 Microcanonical Annealing :1: :8:38 :7:12 Time ::46 :4:19 :2:3 :1:26 :: 294 299 34 39 314 319 324 329 334 339 344 :11:31 Tabu Search :1: :8:38 :7:12 Time ::46 :4:19 :2:3 :1:26 :: 294 299 34 39 314 319 324 329 334 339 344 Fig. 3. Execution times versus final costs obtained in fifty runs for K. Times in minutes. the intra-group paths for posterior processing. In contrast to that, tabu search, by accepting, at each step, the least expensive move (as long as it is not tabu), restricts itself, most of the time, to short-scale changes in the solutions. Therefore, it has difficulty in processing the large-scale corrections of the paths between groups. Similarly, µo finds it hard to evolve in such topology, unless the demon parameters are chosen large enough to accomodate large-scale rearrangements. For this reason, in our implementations for PR124 and PR439, instead of the th entry in the list of rejected moves, we had to choose, for the demon parameters, the th term there. As illustrated in Fig. 4, for PR439, tabu search, which received no special tuning for this particular situation, fared worse in those problems.

142 A. Linhares & J. R. A. Torreão :: Microcanonical Optimization Time :17:17 :8:38 :: 17 19 111 113 1 117 119 121 123 1 :: Microcanonical Annealing :17:17 Time :8:38 :: 17 19 111 113 1 117 119 121 123 1 :: Tabu Search :17:17 Time :8:38 :: 17 19 111 113 1 117 119 121 123 1 Fig. 4. Execution times versus final costs obtained in fifty runs for PR439. Times in minutes. It is interesting, in this respect, to remark that µo seems to be more efficient than tabu search in breaking loose from local minimum configurations. The curves in Fig., obtained for problem P3, illustrate this. The plots show the values of the current solution and of the best solution so far encountered, as the algorithms evolve. The tabu heuristic, once in a local minimum, accepts the best of the proposed moves, irrespective of its cost. Since moves which are quite bad can thus be accepted repeatedly, the heuristic tends to stray from the best solution so far obtained. This should be compared to the behavior of µo, where the limited capacity of the demon keeps the current and the best solutions always close. This, nevertheless, does not seem to compromise the quality of the overall optimization: the algorithm is able

Microcanonical Optimization Applied to the Traveling Salesman Problem 143 3 298 296 294 29 29 288 286 284 28 28 3 298 296 294 29 29 288 286 284 28 28 Tabu Search 1 6 11 16 21 26 31 36 41 46 1 6 61 66 71 76 81 Steps Microcanonical Optimization 1 6 11 16 21 26 31 36 41 46 1 6 61 66 71 76 81 Steps Fig.. Comparative evolution of current solution (fine line) and best solution (thick line), at each implementation step, for P3. Microcanonical Optimization 1 28 29 3 317 33 Tabu Search 1 28 29 37 3 33 Fig. 6., in fifty runs, of the final costs obtained for Problem P3, with execution time limited to 3 min.

144 A. Linhares & J. R. A. Torreão to find a way to a near-optimal solution, passing only through intermediary states which are approximately local minima. Finally, since the quality of the final results is also a function of the execution time, and since µo and tabu search obey different stopping criteria, we also compared their performance in limited-time implementations. The distributions of results obtained in runs for P3, with a time limit of 3 min, are shown in Fig. 6, which makes clear, once again, the better performance of µo.. Conclusions We have presented an analysis of the performance of a new heuristic the microcanonical optimization algorithm, µo when applied to the euclidean traveling salesman problem. When confronted with alternative approaches to the TSP (simulated annealing, microcanonical annealing, Tsallis annealing and tabu search), µo yielded the best overall results in our experiments. We have found it to be consistently faster than simulated annealing and consistently superior, in terms of solution quality, to the Tsallis annealing, even though the latter proved to be an efficient strategy for finding plausible solutions in short running times, as already claimed. 3 Microcanonical annealing and tabu search also performed well in our analysis. Due to the adaptive divide-and-conquer nature of the annealing, MA was able to outperform tabu search (though not µo), in what concerns the quality of the solutions, in certain problems with highly non-uniform city distributions, which require a scale-dependent processing. In most of the other experiments, tabu search proved itself the closest competitor to µo, yielding slightly inferior results in comparable execution times. We conclude that µo is a very promising heuristic for combinatorial optimization problems, as demonstrated by its extremely robust and efficient performance in the benchmark application of the TSP. References 1. V. Cerny, J. Optimization Theory and Applications 4, 41 (198). 2. J. J. Hopfield and D. W. Tank, Bio. Cyber. 2, 141 (198). 3. T. J. P. Penna, Phys. Rev. E1, 1 (199). 4. S. Kirkpatrick, D. C. Gellat, and M. Vecchi, Science 2, 671 (1983).. H. Szu and R. Hartley, Phys. Lett. A122, 7 (1987). 6. S. T. Barnard, Int. J. Comp. Vision 3, 17 (1989). 7. J. R. A. Torreão and E. Roe, Phys. Lett. A, 377 (199). 8. J. L. Fernandes and J. R. A. Torreão, in Lecture Notes in Computer Science Proc. 3rd. Asian Conf. on Computer Vision (Springer-Verlag, Heidelberg, 1998), to appear. 9. S. C. S. Porto, A. M. Barroso, and J. R. A. Torreão, in Proc. 2nd. Methaheuristics Int. Conf. (INRIA, Sophia-Antipolis, 1997), p. 13. 1. F. Glover, ORSA J. Comp. 1, 19 (1989); ORSA J. Comp. 2, 4 (199). 11. M. Creutz, Phys. Rev. Lett., 1411 (1983). 12. N. Metropolis, A. W. Rosenbluth, M. N. Rosenbluth, A. H. Teller, and E. Teller, The J. Chem. Phys. 21, 187 (193).

Microcanonical Optimization Applied to the Traveling Salesman Problem 14 13. L. E. Reichl, A Modern Course on Statistical Physics (The University of Texas Press, Austin, 1986). 14. P. J. M. Laarhoven and E. H. L. Aarts, Simulated Annealing: Theory and Applications (Kluwer Academic Publishers, Amsterdam, 1987).. W. H. Press, S. A. Teukolsky, W. T. Vetterling, and B. P. Flannery, Numerical Recipes: The Art of Scientific Computing (Cambridge University Press, Cambridge, 1992). 16. C. Tsallis, J. Stat. Phys. 2, 479 (1988). 17. S. Lin and B. W. Kernighan, Op. Res. 21, 498 (1973). 18. G. Reinelt, ORSA J. Comp. 3, 376 (1991). Appendix A Here we present the pseudocode for the microcanonical optimization metaheuristic. µo algorithm Let maxcycle be the maximum number of iterations without improvement of the solution cost; repeat do Initialization; do Sampling; until (maxcycle is reached) end Fig. A.1. µo algorithm. procedure Initialization Empty list-of-rejected-moves; Let maxinit be the maximum number of consecutive rejected moves; Let s be the starting solution of the initilization phase; num rejmoves ; while (num rejmoves < maxinit) do Choose a move randomly; Call the new solution s ; Compute cost E of solution s; Compute cost E of solution s ; costchange E E if (costchange ) then Put costchange in the list-of-rejected-moves; num rejmoves num rejmoves +1; end if else num rejmoves s s

146 A. Linhares & J. R. A. Torreão end else end while end Fig. A.2. Initialization procedure. procedure Sampling Select D MAX and D I from the list-of-rejected-moves; Let maxsamp be the maximum number of sampling iterations; Let s be the starting solution of the sampling phase; num iter ; D D I while (num iter < maxsamp) do Choose a move randomly; Call the new solution s ; Compute cost E of solution s; Compute cost E of solution s ; costchange E E if (costchange ) then if (D costchange D MAX ) then s s ; D D costchange; end if end if else {costchange > } if (D costchange ) then s s ; D D costchange; end if end else num iter num iter +1; end while end Fig. A.3. Sampling procedure.