Motivation, Basic Concepts, Basic Methods, Travelling Salesperson Problem (TSP), Algorithms

Size: px
Start display at page:

Download "Motivation, Basic Concepts, Basic Methods, Travelling Salesperson Problem (TSP), Algorithms"

Transcription

1 Motivation, Basic Concepts, Basic Methods, Travelling Salesperson Problem (TSP), Algorithms 1

2 What is Combinatorial Optimization? Combinatorial Optimization deals with problems where we have to search for and find an optimal combination of objects from a very large, but finite, number of possible combinations of objects. For example, choose the best combination of 42 objects from 1,234 objects. The set containing the finite number of possibilities is called the solution space (possibility space). Continuing the example, the solution space consists of all possible combinations of 42 objects taken from the set of 1,234 objects. Each point in the solution space (which is a combination of objects) is a possible solution, and each solution has a cost or benefit associated with it. Finding an optimal combination of objects means finding a combination of objects that minimizes the cost or maximizes the benefit function. Give lowest distance Give highest profit Within the solution space, there may be many solutions that could represent very good solutions (local optimum solutions), which have very low cost or very high benefit, and one best solution (i.e., global optimum solution), which has the lowest cost or the highest benefit. 2

3 Combinatorial Optimization Example You have n = 10,000 objects. Each object has a different value: Object 1: $9.47 Object 2: $5.32 Object n: $7.44 That would take a long time, Hairy! A combination is defined as a subset of 100 objects. Each object has to be used at least once, and not more than once. The solution space consists of all possible combinations of 100 objects ( e+139 combinations) Can you find a combination of 100 objects whose total value is closest to $1,245,678.90? 3

4 CO Example: Travelling Salesman Problem The Travelling Salesman Problem (TSP) describes a salesperson who must travel a route to visit cities. The distance between city and city is known. The salesperson s problem is to: 1. Visit each city exactly once (i.e., visit each city at least once and no more than once) 2. Return to the starting point (city). The starting city (starting point of the route) can be any of the cities. 3. Find a route that represents the minimum distance travelled. Each possible route is an ordered combination of cities to visit. Each possible route (ordered combination) is a possible solution in the solution space. That is a difficult problem! Each route has a cost: the total distance travelled. Objective: find the route that has the least distance. 4

5 Combinatorial Optimization Algorithms Global optimum algorithms: Exhaustive search Held and Karp (1962): dynamic programming Branch-and-cut (later branch and bound) Concorde implementation holds current record (finding best route for 85,900 cities) However, even for the state-of-the-art Concorde implementation, these exact algorithms take a long time to compute. In April 2006 Concorde TSP Solver of 85,900 cities took over 136 CPU-years (A CPU Year is the amount of computing work done by a 1 GFLOP reference machine in a year of dedicated service.) Evolutionary algorithms (Heuristic and approximate) Hill climbing (and variants, such as random restarts) Simulated annealing Genetic algorithm Artificial neural network Ant Colony Optimization Particle Swarm Optimization 5

6 Exhaustive Search suffers from a serious problem as the number of variables increases, the number of combinations to be examined explodes. For example, consider the travelling salesperson problem with a total number of cities to visit equal to 23. There are different possible solutions. Given a starting city, we have n-1 choices for the second city, n-2 choices for the third city, etc. Multiplying these together we get (n-1)! = n-1 x n-2 x n-3 x... x 3 x 2 x 1. Now since our travel costs do not depend on the direction we take around the tour, we should divide this number by 2 to get (n-1)!/2 How large a number is? Suppose we have a computer capable of evaluating a feasible solution in one ns ( s). If we had only 23 cities to visit, then it would take approximately 178 centuries to run through the possible tours. 6

7 In contrast, evolutionary algorithms do not suffer from this problem of taking forever to solve a problem. These algorithms are so-named evolutionary because They mimic natural processes that govern how processes in nature naturally evolve. They are iterative algorithms that evolve incrementally (hopefully improve on their solution) through each iteration. These algorithms do not attempt to examine the entire space. Even so, they have been shown to provide good solutions in a fraction of the amount time, as compared to exact algorithms, which, for large size problems, take a large amount of time, i.e., they are infeasible. 7

8 The annealing of solids is a process in nature that naturally evolves. The term annealing refers to the process in which a solid, that has been brought into liquid phase by increasing its temperature, is brought back to a solid phase by slowly reducing the temperature in such a way that all the particles are allowed to arrange themselves in the strongest possible crystallized state. Such a crystallized state represents the global minimum of the solid s energy function. The cooling process has to be slow enough in order to guarantee that the particles will have time and opportunity to rearrange themselves and find the best position at the current temperature, for all temperature settings in the cooling process. At a certain temperature, once the particles have reached their best position, the substance is said to have reached thermal equilibrium. Once thermal equilibrium is reached, the temperature is lowered once again, and the process continues. At the completion, atoms have attained a nearly globally minimum energy state. 8

9 During the annealing process, the probability of particle s motion is given by the Boltzmann probability function: The probability of a particle s range of motion is proportional to the temperature. At higher temperatures, the particles have a greater range of motion than at lower temperatures. The higher temperature at the beginning of the annealing process allows the particles to move a greater range than at lower temperatures. As the temperature decreases, the particles range of motion decreases exponentially. The probability of a particle s range of motion is inversely proportional to the change of energy from one temperature state to another for large positive changes in energy from one state to another, which means the particle enters a much lower energy state, the probability of movements gets smaller. (Large positive change in energy means finding good solutions, so reduce the movements to look elsewhere.) 9

10 Dependence on Temperature and Cooling Schedule The annealing process consists of first raising the temperature of a solid to a point where its atoms can freely (i.e., randomly) move and then to lower the temperature, forcing the atoms to rearrange themselves into a lower energy state (i.e., a crystallization process). The cooling schedule is vital in this process. If the solid is cooled too quickly, or if the initial temperature of the system is too low, at the end of the annealing process, the substance is not able to become a crystal and instead the solid arrives at an amorphous state with higher energy. In this case, the system reaches a poor local minimum (a higher energy state) instead of the global minimum (i.e., the minimal energy state). For example, if you let metal cool rapidly, its atoms aren t given a chance to settle into a tight lattice and are frozen in a random configuration, resulting in brittle metal. If we decrease the temperature very slowly, the atoms are given enough time to settle into a strong crystal. 10

11 Application of Annealing to Combinatorial Optimization Problems The idea of the annealing process may be applied to combinatorial optimization problems. [1]Metropolis, Nicholas; Rosenbluth, Arianna W.; Rosenbluth, Marshall N.; Teller, Augusta H.; Teller, Edward (1953). "Equation of State Calculations by Fast Computing Machines". The Journal of Chemical Physics 21 (6): Bibcode:1953JChPh M. doi: / [2]Kirkpatrick, S.; Gelatt Jr, C. D.; Vecchi, M. P. (1983). "Optimization by Simulated Annealing". Science 220 (4598): Bibcode:1983Sci K. doi: /science JSTOR PMID

12 Application to Combinatorial Optimization Problems, such as TSP The idea of the annealing process may be applied to combinatorial optimization problems, such as the Traveling Salesperson Problem (TSP). There are many molecules in a substance. Many cities to visit (list of cities) There are much more positions in the substance that a molecule can take. Many possible routes to visit the cities in the list Each set of positions the molecules can take is called a combination. Each route (ordered sequence of cities to visit) is called a combination Each solution (each combination) has an associated cost, which is the amount of the energy represented by the combination. Each route has a cost, which is the total distance of the route The optimization problem is to find a combination that has the lowest cost (energy) Shortest distance. 12

13 Simulated Annealing Algorithm: Basic Structure The basic simulated annealing algorithm can be described as an iterative and evolutionary procedure composed of two loops: an outer loop and a nested inner loop. The inner loop simulates the goal of attaining thermal equilibrium at a given temperature. This is called the Thermal Equilibrium Loop. The outer loop performs the cooling process, in which the temperature is decreased from its initial value towards zero until a certain termination criterion is achieved and the search is stopped. This is called the Cooling Loop. The algorithm starts by initializing several parameters: The initial temperature is set to a very high value (to mimic the temperature setting of the initial natural annealing process). This is needed to allow the algorithm to search a wide breadth of solutions initially. The initial solution is created. Usually, the initial solution is chosen randomly. The number of times to go through the inner loop and the outer loop is set. 13

14 Each time the Thermal Equilibrium Loop (i.e., the inner loop) is called, it is run with a constant temperature, and the goal of the inner loop is to find a best solution for the given temperature to attain thermal equilibrium. Each iteration through the Thermal Equilibrium Loop, the algorithm performs the following: A small random perturbation of the currently held, best-so-far solution is made to create a new candidate solution. Since the algorithm does not know which direction to search, it picks a random direction by randomly perturbing the current solution. The goodness of a solution is quantified by a cost function. For example, the cost function for the TSP is the sum of the distances between each successive city in the solution list of cities to visit. We can think of the solution space by imagining a cost surface in hyperspace, where each location in hyperspace is given by a route (i.e., ordered list of cities), and the height of the cost function at a location represents the cost value. The algorithm wants to go to the minimum point (route) location that represents the lowest cost. A small random perturbation is made to the current solution, thus creating a candidate solution. A small step is taken because it is believed that good solutions are generally close to each other. But, this is not guaranteed to be true all of the time, because it depends on the problem and the distribution of solutions in the solution space. 14

15 Each iteration through the Thermal Equilibrium Loop, continued: Sometimes, the random perturbation results in a better solution: If the cost of the new candidate solution is lower than the cost of the previous solution, i.e., the random perturbation results in a better solution, then the new solution is kept and replaces the previous solution. Sometimes, the random perturbation results in a worse solution, in which case the algorithm makes a decision to keep or discard this worse solution. The decision outcome depends on an evaluation of a probability function, which depends on the temperature of the current loop and the change of error the candidate solution represents. The higher is the temperature, the more likely the algorithm will keep a worse solution. Keeping a worse solution is done to allow the algorithm to explore the solution space and to keep it from being trapped in a local minima. The lower is the temperature, the less likely the algorithm will keep a worse solution. Discarding a worse solution allows the algorithm to exploit a local optimum, which might be the global optimum. 15

16 Each iteration through the Thermal Equilibrium Loop, continued: The decision outcome depends on an evaluation of an estimation [1] of the Boltzmann's probability function: The Boltzmann s probability fun For a real annealing process, the is the change in energy of the atoms from the previous temperature state and the current temperature state, where the energy is given by the potential and kinetic energy of the atoms in the substance at the given temperature of the state. For simulated annealing, this may be estimated by the change in the cost function corresponding to the difference between cost of the previously found best solution at its temperature state and the cost of the new candidate solution at the current temperature state. The Boltzmann's constant may be estimated by the average cost function taken over all iterations of the inner loop. Therefore, 16

17 Each iteration through the Thermal Equilibrium Loop, continued: The decision outcome depends on: As can been seen, this probability is proportional to the temperature and inversely proportional to the normalized change in cost of the current solution. 1 17

18 Each iteration through the Thermal Equilibrium Loop, continued: Thus, in one iteration of the inner loop, the algorithm will either find and keep a better solution, keep a worse solution with probability, or make no change and keep the previously found best solution. The algorithm continues running the inner loop and the above procedure for a number of times. After running the inner loop many times, where in each loop it takes on a new better solution, or takes on a worse solution, or keeps the previously found best solution, the algorithm may be viewed as taking a random walk in the solution space, looking for a stable sub-optimal solution for the given temperature of the inner loop. After having found a stable sub-optimal solution for the given temperature of the inner loop, the process is said to have reached thermal equilibrium. At this point the inner loop completes, and the algorithm goes to the outer loop. 18

19 The outer loop (i.e., the Cooling Loop) performs the following: The currently best solution is recorded as the optimal solution. The temperature is decreased, according to some schedule. The initial temperature is set to a very high value (to mimic the temperature setting of the initial natural annealing process). This is needed to allow the algorithm to search a wide breadth of solutions initially. The final temperature should be set to some low value to prevent the algorithm from accepting worse solutions at the late stages of the process. The number of outer loops is decremented. If the number outer loops hasn t reached zero, then the inner loop is called once again; otherwise the algorithm terminates. 19

20 Set #ncl, itemp. Done Get Initial Route Y #ncl s==0? Dec # ncl N Compute Distance of Initial Route Set #nel perturbroute N Equilibrium Loop: The temperature is held constant, while the system reaches equilibrium, i.e., until the best route if found for the given temperature. Compute Distance of Route Reduce Temperature Y #nel ==0? Dec # nel Y? Better Route? ncl=numcoolingloops nel=numequilibriumloops N? Y genrand # N Find Prob of Acceptance, 20

21 21

22 Watch YouTube Simulation 22

23 [1] J. D. Hedengren, "Optimization Techniques in Engineering," 5 April [Online]. Available: [Accessed 27 April 2015]. [2] A. R. Parkinson, R. J. Balling and J. D. Heden, "Optimization Methods for Engineering Design Applications and Theory," Brigham Young University,

5. Simulated Annealing 5.1 Basic Concepts. Fall 2010 Instructor: Dr. Masoud Yaghini

5. Simulated Annealing 5.1 Basic Concepts. Fall 2010 Instructor: Dr. Masoud Yaghini 5. Simulated Annealing 5.1 Basic Concepts Fall 2010 Instructor: Dr. Masoud Yaghini Outline Introduction Real Annealing and Simulated Annealing Metropolis Algorithm Template of SA A Simple Example References

More information

Motivation, Basic Concepts, Basic Methods, Travelling Salesperson Problem (TSP), Algorithms

Motivation, Basic Concepts, Basic Methods, Travelling Salesperson Problem (TSP), Algorithms Motivation, Basic Concepts, Basic Methods, Travelling Salesperson Problem (TSP), Algorithms 1 Next Step Application of simulated annealing to the Traveling Salesman problem. 2 Choosing a temperature schedule.

More information

Methods for finding optimal configurations

Methods for finding optimal configurations CS 1571 Introduction to AI Lecture 9 Methods for finding optimal configurations Milos Hauskrecht milos@cs.pitt.edu 5329 Sennott Square Search for the optimal configuration Optimal configuration search:

More information

7.1 Basis for Boltzmann machine. 7. Boltzmann machines

7.1 Basis for Boltzmann machine. 7. Boltzmann machines 7. Boltzmann machines this section we will become acquainted with classical Boltzmann machines which can be seen obsolete being rarely applied in neurocomputing. It is interesting, after all, because is

More information

SIMU L TED ATED ANNEA L NG ING

SIMU L TED ATED ANNEA L NG ING SIMULATED ANNEALING Fundamental Concept Motivation by an analogy to the statistical mechanics of annealing in solids. => to coerce a solid (i.e., in a poor, unordered state) into a low energy thermodynamic

More information

A.I.: Beyond Classical Search

A.I.: Beyond Classical Search A.I.: Beyond Classical Search Random Sampling Trivial Algorithms Generate a state randomly Random Walk Randomly pick a neighbor of the current state Both algorithms asymptotically complete. Overview Previously

More information

( ) ( ) ( ) ( ) Simulated Annealing. Introduction. Pseudotemperature, Free Energy and Entropy. A Short Detour into Statistical Mechanics.

( ) ( ) ( ) ( ) Simulated Annealing. Introduction. Pseudotemperature, Free Energy and Entropy. A Short Detour into Statistical Mechanics. Aims Reference Keywords Plan Simulated Annealing to obtain a mathematical framework for stochastic machines to study simulated annealing Parts of chapter of Haykin, S., Neural Networks: A Comprehensive

More information

Methods for finding optimal configurations

Methods for finding optimal configurations S 2710 oundations of I Lecture 7 Methods for finding optimal configurations Milos Hauskrecht milos@pitt.edu 5329 Sennott Square S 2710 oundations of I Search for the optimal configuration onstrain satisfaction

More information

Finding optimal configurations ( combinatorial optimization)

Finding optimal configurations ( combinatorial optimization) CS 1571 Introduction to AI Lecture 10 Finding optimal configurations ( combinatorial optimization) Milos Hauskrecht milos@cs.pitt.edu 539 Sennott Square Constraint satisfaction problem (CSP) Constraint

More information

Optimization Methods via Simulation

Optimization Methods via Simulation Optimization Methods via Simulation Optimization problems are very important in science, engineering, industry,. Examples: Traveling salesman problem Circuit-board design Car-Parrinello ab initio MD Protein

More information

PROBLEM SOLVING AND SEARCH IN ARTIFICIAL INTELLIGENCE

PROBLEM SOLVING AND SEARCH IN ARTIFICIAL INTELLIGENCE Artificial Intelligence, Computational Logic PROBLEM SOLVING AND SEARCH IN ARTIFICIAL INTELLIGENCE Lecture 4 Metaheuristic Algorithms Sarah Gaggl Dresden, 5th May 2017 Agenda 1 Introduction 2 Constraint

More information

Introduction to Simulated Annealing 22c:145

Introduction to Simulated Annealing 22c:145 Introduction to Simulated Annealing 22c:145 Simulated Annealing Motivated by the physical annealing process Material is heated and slowly cooled into a uniform structure Simulated annealing mimics this

More information

Heuristic Optimisation

Heuristic Optimisation Heuristic Optimisation Part 8: Simulated annealing Sándor Zoltán Németh http://web.mat.bham.ac.uk/s.z.nemeth s.nemeth@bham.ac.uk University of Birmingham S Z Németh (s.nemeth@bham.ac.uk) Heuristic Optimisation

More information

Simulated Annealing. Local Search. Cost function. Solution space

Simulated Annealing. Local Search. Cost function. Solution space Simulated Annealing Hill climbing Simulated Annealing Local Search Cost function? Solution space Annealing Annealing is a thermal process for obtaining low energy states of a solid in a heat bath. The

More information

Hill climbing: Simulated annealing and Tabu search

Hill climbing: Simulated annealing and Tabu search Hill climbing: Simulated annealing and Tabu search Heuristic algorithms Giovanni Righini University of Milan Department of Computer Science (Crema) Hill climbing Instead of repeating local search, it is

More information

Local Search & Optimization

Local Search & Optimization Local Search & Optimization CE417: Introduction to Artificial Intelligence Sharif University of Technology Spring 2017 Soleymani Artificial Intelligence: A Modern Approach, 3 rd Edition, Chapter 4 Outline

More information

Single Solution-based Metaheuristics

Single Solution-based Metaheuristics Parallel Cooperative Optimization Research Group Single Solution-based Metaheuristics E-G. Talbi Laboratoire d Informatique Fondamentale de Lille Single solution-based metaheuristics Improvement of a solution.

More information

Zebo Peng Embedded Systems Laboratory IDA, Linköping University

Zebo Peng Embedded Systems Laboratory IDA, Linköping University TDTS 01 Lecture 8 Optimization Heuristics for Synthesis Zebo Peng Embedded Systems Laboratory IDA, Linköping University Lecture 8 Optimization problems Heuristic techniques Simulated annealing Genetic

More information

Ant Colony Optimization: an introduction. Daniel Chivilikhin

Ant Colony Optimization: an introduction. Daniel Chivilikhin Ant Colony Optimization: an introduction Daniel Chivilikhin 03.04.2013 Outline 1. Biological inspiration of ACO 2. Solving NP-hard combinatorial problems 3. The ACO metaheuristic 4. ACO for the Traveling

More information

Design and Analysis of Algorithms

Design and Analysis of Algorithms CSE 0, Winter 08 Design and Analysis of Algorithms Lecture 8: Consolidation # (DP, Greed, NP-C, Flow) Class URL: http://vlsicad.ucsd.edu/courses/cse0-w8/ Followup on IGO, Annealing Iterative Global Optimization

More information

Artificial Intelligence Heuristic Search Methods

Artificial Intelligence Heuristic Search Methods Artificial Intelligence Heuristic Search Methods Chung-Ang University, Jaesung Lee The original version of this content is created by School of Mathematics, University of Birmingham professor Sandor Zoltan

More information

Discrete evaluation and the particle swarm algorithm

Discrete evaluation and the particle swarm algorithm Volume 12 Discrete evaluation and the particle swarm algorithm Tim Hendtlass and Tom Rodgers Centre for Intelligent Systems and Complex Processes Swinburne University of Technology P. O. Box 218 Hawthorn

More information

Computational statistics

Computational statistics Computational statistics Combinatorial optimization Thierry Denœux February 2017 Thierry Denœux Computational statistics February 2017 1 / 37 Combinatorial optimization Assume we seek the maximum of f

More information

Overview. Optimization. Easy optimization problems. Monte Carlo for Optimization. 1. Survey MC ideas for optimization: (a) Multistart

Overview. Optimization. Easy optimization problems. Monte Carlo for Optimization. 1. Survey MC ideas for optimization: (a) Multistart Monte Carlo for Optimization Overview 1 Survey MC ideas for optimization: (a) Multistart Art Owen, Lingyu Chen, Jorge Picazo (b) Stochastic approximation (c) Simulated annealing Stanford University Intel

More information

Metaheuristics and Local Search

Metaheuristics and Local Search Metaheuristics and Local Search 8000 Discrete optimization problems Variables x 1,..., x n. Variable domains D 1,..., D n, with D j Z. Constraints C 1,..., C m, with C i D 1 D n. Objective function f :

More information

Artificial Intelligence Methods (G5BAIM) - Examination

Artificial Intelligence Methods (G5BAIM) - Examination Question 1 a) According to John Koza there are five stages when planning to solve a problem using a genetic program. What are they? Give a short description of each. (b) How could you cope with division

More information

Metaheuristics and Local Search. Discrete optimization problems. Solution approaches

Metaheuristics and Local Search. Discrete optimization problems. Solution approaches Discrete Mathematics for Bioinformatics WS 07/08, G. W. Klau, 31. Januar 2008, 11:55 1 Metaheuristics and Local Search Discrete optimization problems Variables x 1,...,x n. Variable domains D 1,...,D n,

More information

1 Heuristics for the Traveling Salesman Problem

1 Heuristics for the Traveling Salesman Problem Praktikum Algorithmen-Entwurf (Teil 9) 09.12.2013 1 1 Heuristics for the Traveling Salesman Problem We consider the following problem. We want to visit all the nodes of a graph as fast as possible, visiting

More information

Data Structures in Java

Data Structures in Java Data Structures in Java Lecture 21: Introduction to NP-Completeness 12/9/2015 Daniel Bauer Algorithms and Problem Solving Purpose of algorithms: find solutions to problems. Data Structures provide ways

More information

Algorithms and Complexity theory

Algorithms and Complexity theory Algorithms and Complexity theory Thibaut Barthelemy Some slides kindly provided by Fabien Tricoire University of Vienna WS 2014 Outline 1 Algorithms Overview How to write an algorithm 2 Complexity theory

More information

12. LOCAL SEARCH. gradient descent Metropolis algorithm Hopfield neural networks maximum cut Nash equilibria

12. LOCAL SEARCH. gradient descent Metropolis algorithm Hopfield neural networks maximum cut Nash equilibria Coping With NP-hardness Q. Suppose I need to solve an NP-hard problem. What should I do? A. Theory says you re unlikely to find poly-time algorithm. Must sacrifice one of three desired features. Solve

More information

Local Search & Optimization

Local Search & Optimization Local Search & Optimization CE417: Introduction to Artificial Intelligence Sharif University of Technology Spring 2018 Soleymani Artificial Intelligence: A Modern Approach, 3 rd Edition, Chapter 4 Some

More information

Fundamentals of Metaheuristics

Fundamentals of Metaheuristics Fundamentals of Metaheuristics Part I - Basic concepts and Single-State Methods A seminar for Neural Networks Simone Scardapane Academic year 2012-2013 ABOUT THIS SEMINAR The seminar is divided in three

More information

22c:145 Artificial Intelligence

22c:145 Artificial Intelligence 22c:145 Artificial Intelligence Fall 2005 Informed Search and Exploration III Cesare Tinelli The University of Iowa Copyright 2001-05 Cesare Tinelli and Hantao Zhang. a a These notes are copyrighted material

More information

Lecture 4: Simulated Annealing. An Introduction to Meta-Heuristics, Produced by Qiangfu Zhao (Since 2012), All rights reserved

Lecture 4: Simulated Annealing. An Introduction to Meta-Heuristics, Produced by Qiangfu Zhao (Since 2012), All rights reserved Lecture 4: Simulated Annealing Lec04/1 Neighborhood based search Step 1: s=s0; Step 2: Stop if terminating condition satisfied; Step 3: Generate N(s); Step 4: s =FindBetterSolution(N(s)); Step 5: s=s ;

More information

Local Search and Optimization

Local Search and Optimization Local Search and Optimization Outline Local search techniques and optimization Hill-climbing Gradient methods Simulated annealing Genetic algorithms Issues with local search Local search and optimization

More information

CS 781 Lecture 9 March 10, 2011 Topics: Local Search and Optimization Metropolis Algorithm Greedy Optimization Hopfield Networks Max Cut Problem Nash

CS 781 Lecture 9 March 10, 2011 Topics: Local Search and Optimization Metropolis Algorithm Greedy Optimization Hopfield Networks Max Cut Problem Nash CS 781 Lecture 9 March 10, 2011 Topics: Local Search and Optimization Metropolis Algorithm Greedy Optimization Hopfield Networks Max Cut Problem Nash Equilibrium Price of Stability Coping With NP-Hardness

More information

5. Simulated Annealing 5.2 Advanced Concepts. Fall 2010 Instructor: Dr. Masoud Yaghini

5. Simulated Annealing 5.2 Advanced Concepts. Fall 2010 Instructor: Dr. Masoud Yaghini 5. Simulated Annealing 5.2 Advanced Concepts Fall 2010 Instructor: Dr. Masoud Yaghini Outline Acceptance Function Initial Temperature Equilibrium State Cooling Schedule Stopping Condition Handling Constraints

More information

Local Search. Shin Yoo CS492D, Fall 2015, School of Computing, KAIST

Local Search. Shin Yoo CS492D, Fall 2015, School of Computing, KAIST Local Search Shin Yoo CS492D, Fall 2015, School of Computing, KAIST If your problem forms a fitness landscape, what is optimisation? Local Search Loop Local Search Loop Start with a single, random solution

More information

Computational Intelligence in Product-line Optimization

Computational Intelligence in Product-line Optimization Computational Intelligence in Product-line Optimization Simulations and Applications Peter Kurz peter.kurz@tns-global.com June 2017 Restricted use Restricted use Computational Intelligence in Product-line

More information

Artificial Intelligence Methods (G5BAIM) - Examination

Artificial Intelligence Methods (G5BAIM) - Examination Question 1 a) Explain the difference between a genotypic representation and a phenotypic representation. Give an example of each. (b) Outline the similarities and differences between Genetic Algorithms

More information

Analysis of a Large Structure/Biological Activity. Data Set Using Recursive Partitioning and. Simulated Annealing

Analysis of a Large Structure/Biological Activity. Data Set Using Recursive Partitioning and. Simulated Annealing Analysis of a Large Structure/Biological Activity Data Set Using Recursive Partitioning and Simulated Annealing Student: Ke Zhang MBMA Committee: Dr. Charles E. Smith (Chair) Dr. Jacqueline M. Hughes-Oliver

More information

Part B" Ants (Natural and Artificial)! Langton s Vants" (Virtual Ants)! Vants! Example! Time Reversibility!

Part B Ants (Natural and Artificial)! Langton s Vants (Virtual Ants)! Vants! Example! Time Reversibility! Part B" Ants (Natural and Artificial)! Langton s Vants" (Virtual Ants)! 11/14/08! 1! 11/14/08! 2! Vants!! Square grid!! Squares can be black or white!! Vants can face N, S, E, W!! Behavioral rule:!! take

More information

Local and Stochastic Search

Local and Stochastic Search RN, Chapter 4.3 4.4; 7.6 Local and Stochastic Search Some material based on D Lin, B Selman 1 Search Overview Introduction to Search Blind Search Techniques Heuristic Search Techniques Constraint Satisfaction

More information

CS 380: ARTIFICIAL INTELLIGENCE

CS 380: ARTIFICIAL INTELLIGENCE CS 380: ARTIFICIAL INTELLIGENCE PROBLEM SOLVING: LOCAL SEARCH 10/11/2013 Santiago Ontañón santi@cs.drexel.edu https://www.cs.drexel.edu/~santi/teaching/2013/cs380/intro.html Recall: Problem Solving Idea:

More information

How hard is it to find a good solution?

How hard is it to find a good solution? How hard is it to find a good solution? Simons Institute Open Lecture November 4, 2013 Research Area: Complexity Theory Given a computational problem, find an efficient algorithm that solves it. Goal of

More information

MCMC Simulated Annealing Exercises.

MCMC Simulated Annealing Exercises. Aula 10. Simulated Annealing. 0 MCMC Simulated Annealing Exercises. Anatoli Iambartsev IME-USP Aula 10. Simulated Annealing. 1 [Wiki] Salesman problem. The travelling salesman problem (TSP), or, in recent

More information

Unit 1A: Computational Complexity

Unit 1A: Computational Complexity Unit 1A: Computational Complexity Course contents: Computational complexity NP-completeness Algorithmic Paradigms Readings Chapters 3, 4, and 5 Unit 1A 1 O: Upper Bounding Function Def: f(n)= O(g(n)) if

More information

SYSTEMS SCIENCE AND CYBERNETICS Vol. III Simulated Annealing: From Statistical Thermodynamics to Combinatory Problems Solving - D.

SYSTEMS SCIENCE AND CYBERNETICS Vol. III Simulated Annealing: From Statistical Thermodynamics to Combinatory Problems Solving - D. SIMULATED ANNEALING: FROM STATISTICAL THERMODYNAMICS TO COMBINATORY PROBLEMS SOLVING D. Thiel ENITIAA, Nantes, France Keywords: Combinatory Problem, Optimizing Problem, Global Search Method, Statistical

More information

Local search algorithms. Chapter 4, Sections 3 4 1

Local search algorithms. Chapter 4, Sections 3 4 1 Local search algorithms Chapter 4, Sections 3 4 Chapter 4, Sections 3 4 1 Outline Hill-climbing Simulated annealing Genetic algorithms (briefly) Local search in continuous spaces (very briefly) Chapter

More information

Lecture H2. Heuristic Methods: Iterated Local Search, Simulated Annealing and Tabu Search. Saeed Bastani

Lecture H2. Heuristic Methods: Iterated Local Search, Simulated Annealing and Tabu Search. Saeed Bastani Simulation Lecture H2 Heuristic Methods: Iterated Local Search, Simulated Annealing and Tabu Search Saeed Bastani saeed.bastani@eit.lth.se Spring 2017 Thanks to Prof. Arne Løkketangen at Molde University

More information

Week Cuts, Branch & Bound, and Lagrangean Relaxation

Week Cuts, Branch & Bound, and Lagrangean Relaxation Week 11 1 Integer Linear Programming This week we will discuss solution methods for solving integer linear programming problems. I will skip the part on complexity theory, Section 11.8, although this is

More information

Random Search. Shin Yoo CS454, Autumn 2017, School of Computing, KAIST

Random Search. Shin Yoo CS454, Autumn 2017, School of Computing, KAIST Random Search Shin Yoo CS454, Autumn 2017, School of Computing, KAIST Random Search The polar opposite to the deterministic, examineeverything, search. Within the given budget, repeatedly generate a random

More information

Metaheuristics. 2.3 Local Search 2.4 Simulated annealing. Adrian Horga

Metaheuristics. 2.3 Local Search 2.4 Simulated annealing. Adrian Horga Metaheuristics 2.3 Local Search 2.4 Simulated annealing Adrian Horga 1 2.3 Local Search 2 Local Search Other names: Hill climbing Descent Iterative improvement General S-Metaheuristics Old and simple method

More information

Constraint satisfaction search. Combinatorial optimization search.

Constraint satisfaction search. Combinatorial optimization search. CS 1571 Introduction to AI Lecture 8 Constraint satisfaction search. Combinatorial optimization search. Milos Hauskrecht milos@cs.pitt.edu 539 Sennott Square Constraint satisfaction problem (CSP) Objective:

More information

Implementation of Travelling Salesman Problem Using ant Colony Optimization

Implementation of Travelling Salesman Problem Using ant Colony Optimization RESEARCH ARTICLE OPEN ACCESS Implementation of Travelling Salesman Problem Using ant Colony Optimization Gaurav Singh, Rashi Mehta, Sonigoswami, Sapna Katiyar* ABES Institute of Technology, NH-24, Vay

More information

Pengju

Pengju Introduction to AI Chapter04 Beyond Classical Search Pengju Ren@IAIR Outline Steepest Descent (Hill-climbing) Simulated Annealing Evolutionary Computation Non-deterministic Actions And-OR search Partial

More information

Solving the flow fields in conduits and networks using energy minimization principle with simulated annealing

Solving the flow fields in conduits and networks using energy minimization principle with simulated annealing Solving the flow fields in conduits and networks using energy minimization principle with simulated annealing Taha Sochi University College London, Department of Physics & Astronomy, Gower Street, London,

More information

6. APPLICATION TO THE TRAVELING SALESMAN PROBLEM

6. APPLICATION TO THE TRAVELING SALESMAN PROBLEM 6. Application to the Traveling Salesman Problem 92 6. APPLICATION TO THE TRAVELING SALESMAN PROBLEM The properties that have the most significant influence on the maps constructed by Kohonen s algorithm

More information

Lecture 35 Minimization and maximization of functions. Powell s method in multidimensions Conjugate gradient method. Annealing methods.

Lecture 35 Minimization and maximization of functions. Powell s method in multidimensions Conjugate gradient method. Annealing methods. Lecture 35 Minimization and maximization of functions Powell s method in multidimensions Conjugate gradient method. Annealing methods. We know how to minimize functions in one dimension. If we start at

More information

Spin Glas Dynamics and Stochastic Optimization Schemes. Karl Heinz Hoffmann TU Chemnitz

Spin Glas Dynamics and Stochastic Optimization Schemes. Karl Heinz Hoffmann TU Chemnitz Spin Glas Dynamics and Stochastic Optimization Schemes Karl Heinz Hoffmann TU Chemnitz 1 Spin Glasses spin glass e.g. AuFe AuMn CuMn nobel metal (no spin) transition metal (spin) 0.1-10 at% ferromagnetic

More information

Summary. AIMA sections 4.3,4.4. Hill-climbing Simulated annealing Genetic algorithms (briey) Local search in continuous spaces (very briey)

Summary. AIMA sections 4.3,4.4. Hill-climbing Simulated annealing Genetic algorithms (briey) Local search in continuous spaces (very briey) AIMA sections 4.3,4.4 Summary Hill-climbing Simulated annealing Genetic (briey) in continuous spaces (very briey) Iterative improvement In many optimization problems, path is irrelevant; the goal state

More information

Algorithm Design Strategies V

Algorithm Design Strategies V Algorithm Design Strategies V Joaquim Madeira Version 0.0 October 2016 U. Aveiro, October 2016 1 Overview The 0-1 Knapsack Problem Revisited The Fractional Knapsack Problem Greedy Algorithms Example Coin

More information

Simulated Annealing. 2.1 Introduction

Simulated Annealing. 2.1 Introduction Simulated Annealing 2 This chapter is dedicated to simulated annealing (SA) metaheuristic for optimization. SA is a probabilistic single-solution-based search method inspired by the annealing process in

More information

Algorithms. NP -Complete Problems. Dong Kyue Kim Hanyang University

Algorithms. NP -Complete Problems. Dong Kyue Kim Hanyang University Algorithms NP -Complete Problems Dong Kyue Kim Hanyang University dqkim@hanyang.ac.kr The Class P Definition 13.2 Polynomially bounded An algorithm is said to be polynomially bounded if its worst-case

More information

Markov Chain Monte Carlo. Simulated Annealing.

Markov Chain Monte Carlo. Simulated Annealing. Aula 10. Simulated Annealing. 0 Markov Chain Monte Carlo. Simulated Annealing. Anatoli Iambartsev IME-USP Aula 10. Simulated Annealing. 1 [RC] Stochastic search. General iterative formula for optimizing

More information

Introduction to Mathematical Programming IE406. Lecture 21. Dr. Ted Ralphs

Introduction to Mathematical Programming IE406. Lecture 21. Dr. Ted Ralphs Introduction to Mathematical Programming IE406 Lecture 21 Dr. Ted Ralphs IE406 Lecture 21 1 Reading for This Lecture Bertsimas Sections 10.2, 10.3, 11.1, 11.2 IE406 Lecture 21 2 Branch and Bound Branch

More information

DISTRIBUTION SYSTEM OPTIMISATION

DISTRIBUTION SYSTEM OPTIMISATION Politecnico di Torino Dipartimento di Ingegneria Elettrica DISTRIBUTION SYSTEM OPTIMISATION Prof. Gianfranco Chicco Lecture at the Technical University Gh. Asachi, Iaşi, Romania 26 October 2010 Outline

More information

x 2 i 10 cos(2πx i ). i=1

x 2 i 10 cos(2πx i ). i=1 CHAPTER 2 Optimization Written Exercises 2.1 Consider the problem min, where = 4 + 4 x 2 i 1 cos(2πx i ). i=1 Note that is the Rastrigin function see Section C.1.11. a) What are the independent variables

More information

What Is Air Temperature?

What Is Air Temperature? 2.2 Read What Is Air Temperature? In Learning Set 1, you used a thermometer to measure air temperature. But what exactly was the thermometer measuring? What is different about cold air and warm air that

More information

Hopfield Networks and Boltzmann Machines. Christian Borgelt Artificial Neural Networks and Deep Learning 296

Hopfield Networks and Boltzmann Machines. Christian Borgelt Artificial Neural Networks and Deep Learning 296 Hopfield Networks and Boltzmann Machines Christian Borgelt Artificial Neural Networks and Deep Learning 296 Hopfield Networks A Hopfield network is a neural network with a graph G = (U,C) that satisfies

More information

CADERNOS DO IME Série Estatística

CADERNOS DO IME Série Estatística CADERNOS DO IME Série Estatística Universidade do Estado do Rio de Janeiro - UERJ Rio de Janeiro RJ - Brasil ISSN impresso 1413-9022 / ISSN on-line 2317-4535 - v.36, p.29-37, 2014 AN IF-ITEM-DELETED SENSITIVE

More information

Integer Linear Programming

Integer Linear Programming Integer Linear Programming Solution : cutting planes and Branch and Bound Hugues Talbot Laboratoire CVN April 13, 2018 IP Resolution Gomory s cutting planes Solution branch-and-bound General method Resolution

More information

Simulated Annealing SURFACE. Syracuse University

Simulated Annealing SURFACE. Syracuse University Syracuse University SURFACE Electrical Engineering and Computer Science Technical Reports College of Engineering and Computer Science 6-1992 Simulated Annealing Per Brinch Hansen Syracuse University, School

More information

CHAPTER 4 INTRODUCTION TO DISCRETE VARIABLE OPTIMIZATION

CHAPTER 4 INTRODUCTION TO DISCRETE VARIABLE OPTIMIZATION CHAPTER 4 INTRODUCTION TO DISCRETE VARIABLE OPTIMIZATION. Introduction.. Examples of Discrete Variables One often encounters problems in which design variables must be selected from among a set of discrete

More information

Natural Computing. Lecture 11. Michael Herrmann phone: Informatics Forum /10/2011 ACO II

Natural Computing. Lecture 11. Michael Herrmann phone: Informatics Forum /10/2011 ACO II Natural Computing Lecture 11 Michael Herrmann mherrman@inf.ed.ac.uk phone: 0131 6 517177 Informatics Forum 1.42 25/10/2011 ACO II ACO (in brief) ACO Represent solution space Set parameters, initialize

More information

Local search algorithms. Chapter 4, Sections 3 4 1

Local search algorithms. Chapter 4, Sections 3 4 1 Local search algorithms Chapter 4, Sections 3 4 Chapter 4, Sections 3 4 1 Outline Hill-climbing Simulated annealing Genetic algorithms (briefly) Local search in continuous spaces (very briefly) Chapter

More information

3D HP Protein Folding Problem using Ant Algorithm

3D HP Protein Folding Problem using Ant Algorithm 3D HP Protein Folding Problem using Ant Algorithm Fidanova S. Institute of Parallel Processing BAS 25A Acad. G. Bonchev Str., 1113 Sofia, Bulgaria Phone: +359 2 979 66 42 E-mail: stefka@parallel.bas.bg

More information

Chapter 4 Beyond Classical Search 4.1 Local search algorithms and optimization problems

Chapter 4 Beyond Classical Search 4.1 Local search algorithms and optimization problems Chapter 4 Beyond Classical Search 4.1 Local search algorithms and optimization problems CS4811 - Artificial Intelligence Nilufer Onder Department of Computer Science Michigan Technological University Outline

More information

Outline. Ant Colony Optimization. Outline. Swarm Intelligence DM812 METAHEURISTICS. 1. Ant Colony Optimization Context Inspiration from Nature

Outline. Ant Colony Optimization. Outline. Swarm Intelligence DM812 METAHEURISTICS. 1. Ant Colony Optimization Context Inspiration from Nature DM812 METAHEURISTICS Outline Lecture 8 http://www.aco-metaheuristic.org/ 1. 2. 3. Marco Chiarandini Department of Mathematics and Computer Science University of Southern Denmark, Odense, Denmark

More information

A Two-Stage Simulated Annealing Methodology

A Two-Stage Simulated Annealing Methodology A Two-Stage Simulated Annealing Methodology James M. Varanelli and James P. Cohoon Department of Computer Science University of Virginia Charlottesville, VA 22903 USA Corresponding Author: James P. Cohoon

More information

Integer Linear Programming (ILP)

Integer Linear Programming (ILP) Integer Linear Programming (ILP) Zdeněk Hanzálek, Přemysl Šůcha hanzalek@fel.cvut.cz CTU in Prague March 8, 2017 Z. Hanzálek (CTU) Integer Linear Programming (ILP) March 8, 2017 1 / 43 Table of contents

More information

3.4 Relaxations and bounds

3.4 Relaxations and bounds 3.4 Relaxations and bounds Consider a generic Discrete Optimization problem z = min{c(x) : x X} with an optimal solution x X. In general, the algorithms generate not only a decreasing sequence of upper

More information

Capacitor Placement for Economical Electrical Systems using Ant Colony Search Algorithm

Capacitor Placement for Economical Electrical Systems using Ant Colony Search Algorithm Capacitor Placement for Economical Electrical Systems using Ant Colony Search Algorithm Bharat Solanki Abstract The optimal capacitor placement problem involves determination of the location, number, type

More information

New rubric: AI in the news

New rubric: AI in the news New rubric: AI in the news 3 minutes Headlines this week: Silicon Valley Business Journal: Apple on hiring spreee for AI experts Forbes: Toyota Invests $50 Million In Artificial Intelligence Research For

More information

OPTIMIZATION BY SIMULATED ANNEALING: A NECESSARY AND SUFFICIENT CONDITION FOR CONVERGENCE. Bruce Hajek* University of Illinois at Champaign-Urbana

OPTIMIZATION BY SIMULATED ANNEALING: A NECESSARY AND SUFFICIENT CONDITION FOR CONVERGENCE. Bruce Hajek* University of Illinois at Champaign-Urbana OPTIMIZATION BY SIMULATED ANNEALING: A NECESSARY AND SUFFICIENT CONDITION FOR CONVERGENCE Bruce Hajek* University of Illinois at Champaign-Urbana A Monte Carlo optimization technique called "simulated

More information

Local search and agents

Local search and agents Artificial Intelligence Local search and agents Instructor: Fabrice Popineau [These slides adapted from Stuart Russell, Dan Klein and Pieter Abbeel @ai.berkeley.edu] Local search algorithms In many optimization

More information

Simulated Annealing applied to the Traveling Salesman Problem. Eric Miller

Simulated Annealing applied to the Traveling Salesman Problem. Eric Miller Simulated Annealing applied to the Traveling Salesman Problem Eric Miller INTRODUCTION The traveling salesman problem (abbreviated TSP) presents the task of finding the most efficient route (or tour) through

More information

Sensitive Ant Model for Combinatorial Optimization

Sensitive Ant Model for Combinatorial Optimization Sensitive Ant Model for Combinatorial Optimization CAMELIA CHIRA cchira@cs.ubbcluj.ro D. DUMITRESCU ddumitr@cs.ubbcluj.ro CAMELIA-MIHAELA PINTEA cmpintea@cs.ubbcluj.ro Abstract: A combinatorial optimization

More information

A Survey on Travelling Salesman Problem

A Survey on Travelling Salesman Problem A Survey on Travelling Salesman Problem Sanchit Goyal Department of Computer Science University of North Dakota Grand Forks, North Dakota 58203 sanchitgoyal01@gmail.com Abstract The Travelling Salesman

More information

Vehicle Routing and Scheduling. Martin Savelsbergh The Logistics Institute Georgia Institute of Technology

Vehicle Routing and Scheduling. Martin Savelsbergh The Logistics Institute Georgia Institute of Technology Vehicle Routing and Scheduling Martin Savelsbergh The Logistics Institute Georgia Institute of Technology Vehicle Routing and Scheduling Part II: Algorithmic Enhancements Handling Practical Complexities

More information

Intuitionistic Fuzzy Estimation of the Ant Methodology

Intuitionistic Fuzzy Estimation of the Ant Methodology BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 9, No 2 Sofia 2009 Intuitionistic Fuzzy Estimation of the Ant Methodology S Fidanova, P Marinov Institute of Parallel Processing,

More information

Math 456: Mathematical Modeling. Tuesday, April 9th, 2018

Math 456: Mathematical Modeling. Tuesday, April 9th, 2018 Math 456: Mathematical Modeling Tuesday, April 9th, 2018 The Ergodic theorem Tuesday, April 9th, 2018 Today 1. Asymptotic frequency (or: How to use the stationary distribution to estimate the average amount

More information

Branch-and-Bound. Leo Liberti. LIX, École Polytechnique, France. INF , Lecture p. 1

Branch-and-Bound. Leo Liberti. LIX, École Polytechnique, France. INF , Lecture p. 1 Branch-and-Bound Leo Liberti LIX, École Polytechnique, France INF431 2011, Lecture p. 1 Reminders INF431 2011, Lecture p. 2 Problems Decision problem: a question admitting a YES/NO answer Example HAMILTONIAN

More information

The Traveling Salesman Problem: A Neural Network Perspective. Jean-Yves Potvin

The Traveling Salesman Problem: A Neural Network Perspective. Jean-Yves Potvin 1 The Traveling Salesman Problem: A Neural Network Perspective Jean-Yves Potvin Centre de Recherche sur les Transports Université de Montréal C.P. 6128, Succ. A, Montréal (Québec) Canada H3C 3J7 potvin@iro.umontreal.ca

More information

MICROCANONICAL OPTIMIZATION APPLIED TO THE TRAVELING SALESMAN PROBLEM

MICROCANONICAL OPTIMIZATION APPLIED TO THE TRAVELING SALESMAN PROBLEM International Journal of Modern Physics C, Vol. 9, No. 1 (1998) 133 146 c World Scientific Publishing Company MICROCANONICAL OPTIMIZATION APPLIED TO THE TRAVELING SALESMAN PROBLEM ALEXANDRE LINHARES Computação

More information

Tabu Search. Biological inspiration is memory the ability to use past experiences to improve current decision making.

Tabu Search. Biological inspiration is memory the ability to use past experiences to improve current decision making. Tabu Search Developed by Fred Glover in the 1970 s. Dr Glover is a business professor at University of Colorado at Boulder. Developed specifically as a combinatorial optimization tool. Biological inspiration

More information

Problem Complexity Classes

Problem Complexity Classes Problem Complexity Classes P, NP, NP-Completeness and Complexity of Approximation Joshua Knowles School of Computer Science The University of Manchester COMP60342 - Week 2 2.15, March 20th 2015 In This

More information

The Traveling Salesman Problem: An Overview. David P. Williamson, Cornell University Ebay Research January 21, 2014

The Traveling Salesman Problem: An Overview. David P. Williamson, Cornell University Ebay Research January 21, 2014 The Traveling Salesman Problem: An Overview David P. Williamson, Cornell University Ebay Research January 21, 2014 (Cook 2012) A highly readable introduction Some terminology (imprecise) Problem Traditional

More information

21. Set cover and TSP

21. Set cover and TSP CS/ECE/ISyE 524 Introduction to Optimization Spring 2017 18 21. Set cover and TSP ˆ Set covering ˆ Cutting problems and column generation ˆ Traveling salesman problem Laurent Lessard (www.laurentlessard.com)

More information