Evolutionary Computation

Size: px
Start display at page:

Download "Evolutionary Computation"

Transcription

1 Evolutionary Computation Lecture Algorithm Configura4on and Theore4cal Analysis

2 Outline Algorithm Configuration Theoretical Analysis 2

3 Algorithm Configuration Question: If an EA toolbox is available (which is true), which operators/parameters shall I use? Aim: To find an appropriate EA instance (configuration of EA) for an optimization task. Representa4on Representa4on 2 Representa4on 3 EA instance Toolbox Operator Operator 2 Operator 3 Operator & Parameter Scheme & Parameter Representa4on Scheme Scheme 2 Scheme 3 3

4 Algorithm Configuration Configuring an EA involves setting the values of categorical variables (type of operators/schemes) numerical variables (e.g., crossover rate, population-size) We refer both the categorical and numerical variables as parameter of EAs More formally, Configuring an EA is another search problem, for which the best parameter vectors are searched to maximize the performance. 4

5 Algorithm Configuration Performance measures may differ in different context: The best solution quality The time required to reach a threshold of solution quality Due to the stochastic nature of EAs, a performance measure can be viewed as a random variable. Multiple runs are usually required to get a reliable estimate of performance. 5

6 Algorithm Configuration The rough idea: Generate-and-Test Example -: A naïve approach Generate p candidate parameter vectors Get p EA instances, each configured with a parameter vector Run each EA instance for r times In each run, n solution vectors is generated and tested Compare the parameter vectors using performance measures to get the best configuration. 6

7 Algorithm Configuration The naïve approach requires calculating the objective function f(x) for prn times. sounds too costly. Smarter approach is needed to reduce search effort. Reduce p Reduce r Reduce n 7

8 Algorithm Configuration Racing procedures: An approach for reducing r. Basic idea: Stop running (testing) an EA instance if there is indication that it is not promising. By not promising, we mean an EA instance performs significantly worse than the best EA instance identified so far. EA instances are compared after each run rather than after running all of them for r times. 8

9 Algorithm Configuration Example -2:. Generate p candidate parameter vectors 2. Set E as the set of p EA instances 3. REPEAT: Run each EA instance once and update its (average) performance Identify the EA instance with the best performance Remove from E the EA instances whose performance is significantly worse than the best EA instance 4. Until size(e)= or the number of runs reaches r Many parameter vectors may be removed without running r times. 9

10 Algorithm Configuration There are a few statistical tests that can be used to indicate significant difference in performance. Analysis of Variance (ANOVA). Kruskal-Wallis (Rank Based Analysis of Variance). Hoeffding s bound. Unpaired Student T-Test. 0

11 Algorithm Configuration Racing procedure may not be sufficient in case: p is very large The set of candidate parameter vectors is infinite Basic idea: To apply an iterative search procedure to the parameter space. Meta-EA: using EAs to search for the best configuration Iterative local search Hopefully, the best configuration can be found with less trial parameter vectors.

12 Algorithm Configuration Example -3:. Generate p candidate parameter vectors (p <p) 2. Get p EA instances 3. REPEAT Run each EA instance for r times and assess its performance Generate another p parameter vectors 4. Until halting criterion is satisfied 2

13 Algorithm Configuration Iterative search can be combined with a racing procedure to reduce both p and r. Example -4:. Generate p candidate parameter vectors (p <p) 2. Get p EA instances 3. REPEAT Applying Racing procedure to compare performance of EA instances Generate another p parameter vectors 4. Until halting criterion is satisfied 3

14 Algorithm Configuration Further enhancement: Modeling the landscape of performance measures For identifying promising parameter vectors For filtering out unpromising parameter vectors The modeling task is a supervised learning problem Input data: parameter vectors Target: performance 4

15 Algorithm Configuration Example -5 (Sequential Parameter Optimization):. Generate p candidate parameter vectors (p <p) 2. Get p EA instances 3. Run each EA instance for r times and assess its performance 4. Build a model M to approximate the performance landscape 5. REPEAT Generate another p parameter vectors Identify the most promising parameter vectors with M Only the promising parameter vectors are run for r times Update M with the newly assessed parameter vectors 6. Until halting criterion is satisfied 5

16 Algorithm Configuration Remark on reducing n The key question involved: Will the relative order of two EAs maintain unchanged during the evolution? SaNSDE SaDE NSDE 0 0 Yes: Blue curve vs. Green Curve No: Blue Curve vs. Red Curve f 9 The lesson learned: the answer is problem and algorithm dependent, and thus few systematic approach have been investigated for this issue. 6

17 Algorithm Configuration So far, we mainly touched the case of configuring EA on a single optimization task (problem instance) What if we want to find a more general configuration that can be applied to multiple problem instances? Previously introduced ideas can be adapted to this scenario. Instead of looking at the results of r runs, comparisons are made mainly based on the performance on multiple problem instances. F-RACE ParamILS 7

18 Algorithm Configuration Additional Remarks Operators/parameters can also be adjusted on-the-fly. This type of EAs are usually called adaptive/self-adaptive EAs. With a unified representation of EA s behaviors, there will huge room for involving machine learning techniques. Algorithm configuration is also closely related to the emerging terminology hyper-heuristic. 8

19 Outline Algorithm Configuration Theoretical Analysis 本部分内容取自南京大学俞扬博士 IJCAI 203 报告的 PPT, 特此表示感谢 9

20 Conventional algorithm analysis Problem Algorithm Measurement Sorting Quick Sort average time complexity O(n log n) Shortest Path Dijkstra s algorithm average time complexity O( V 2 ) Linear Programming Simplex worst case time complexity: exponential smoothed complexity: polynomial 20

21 Time complexity What about an algorithm sorts (5,4,2,8,9) in 3 steps? measured in a class of problem instances e.g. all possible arrays of 5 numbers average complexity worst case complexity measure the growing rate as the problem size increases e.g. 2n 2 asymptotical notation O(n 2 ) 2

22 But for EAs... nature phenomena Problem Algorithm Measurement problem unknown not designed with knowledge of problems theoretical understanding is even more important 22

23 Markov chain modeling A general procedure: initialization evaluation & selection population offspring expand along time: reproduction population 0 population population 2 population 3... Markov chain: state 0 state state 2 state

24 Markov chain: state 0 state state 2 state solution space S P ( 3 2,, 0 )=P ( 3 2 ) optimal solutions S Markov property population space X = S m involves at least one optimal solution optimal populations X 24

25 Convergence Does an EA converge to the global optimal solutions? Considered as closed: lim P ( t 2 X )= t!+ gain of optimality in one step loss of optimality in one step Theorem: (discrete version derived from [He & Yu, 0]) Let be a Markov chain. Define t = X x/2x P ( t+ 2 X t = x)p ( t = x) X x2x P ( t+ /2 X t = x)p ( t = x). Then converges to X if and only if satisfies: P ( 0 2 X )+ +X t=0 t = 25

26 Convergence Does an EA converge to the global optimal solutions? Considered as closed: lim P ( t 2 X )= t!+ An EA that. uses global operators 2. preserves the best solution always converges to the optimal solutions gain of optimality > 0 loss of optimality = 0 But life is limited! How fast does it converge? 26

27 Examples in simple cases Problem OneMax Algorithm Measurement Linear Pseudo-Boolean Functions LeadingOnes LongPath (+)-EA Expected Running Time (ERT) 27

28 Running time analysis Running time of an EA: the number of solutions evaluated until reaching an optimal solution of the given problem for the first time the most time consuming step may meet many times Running time analysis: running time with respect to the problem size (e.g. n) the expected running time/ert e.g. O(n 2 ) expected running time ERT with high probability e.g. computational complexity O(n ln n) expected running time with probability at least 2 n 28

29 A simple EA: (+)-EA An extremely simplified EA missing some features of real EAs (+)-EA no population : s a randomly drawn solution from X 2: for t=,2,... do 3: s 0 mutate(s) 4: if f(s 0 ) f(s) then 5: s s 0 6: end if 7: terminate if meets a stopping criterion 8: end for one-bit mutation randomly choose one bit and change its value bitwise mutation change every bit with some probability (e.g. ) n no crossover for maximization, allow neutral changes find an optimal solution 29

30 OneMax Problem: fitness: arg max x2{0,} n f(x) = nx i= nx i= x i x i count the number of bits EAs do not have the knowledge of the problems only able to call f (x) no difference with any other functions f : {0, } n! R not only optimizing the problem, but also guessing the problem 30

31 probability Theoretical Analysis OneMax: f(x) = nx i= x i (+)-EA with bitwise mutation (flip each bit with probability ): the probability of flipping i particular bits: ( n )i ( monotonically decreasing n )n i n but always positive i 3

32 OneMax: f(x) = nx i= x i the the same number of -bits share the same f value 0 -bits S 0 -bits 2 -bits... n -bits S S 2 S m = S many transitions (+)-EA with bitwise mutation (flip each bit with probability ): n 32

33 OneMax: f(x) = nx i= x i the the same number of -bits share the same f value 0 -bits S 0 -bits 2 -bits... n -bits S S 2 S m = S an upper bound: a path visits all subspaces (+)-EA with bitwise mutation (flip each bit with probability ): n 33

34 OneMax: f(x) = nx i= x i the the same number of -bits share the same f value 0 -bits S 0 -bits 2 -bits... n -bits S S 2 S m = S p = n( n )(n n )n (+)-EA with bitwise mutation (flip each bit with probability ): n 34

35 OneMax: f(x) = nx i= x i the the same number of -bits share the same f value 0 -bits S 0 -bits 2 -bits... n -bits S S 2 S m = S p = n( n )(n p n )n n ( n )(n n )n (+)-EA with bitwise mutation (flip each bit with probability ): n 35

36 OneMax: f(x) = nx i= x i the the same number of -bits share the same f value 0 -bits S 0 -bits 2 -bits... n -bits S S 2 S m = S p = n( n )(n n )n n p ( n )(n n )n p n i ( n )(n n )n (+)-EA with bitwise mutation (flip each bit with probability ): n 36

37 OneMax: f(x) = nx i= x i (+)-EA with bitwise mutation (flip each bit with probability ): n n i probability of transition p ( n )(n n )n expected #steps the transition happens apple n i n ( + n )n n i n e summed up nx i=0 en i = enh n en ln n ERT upper bound O(n ln n) 37

38 General analysis tools running time analysis is commonly problem specific going to derive the ERT of an EA in a problem where to look what to calculate need a guide to tell what to look and what to follow to accomplish the analysis - Fitness Level Method - Drift Analysis - 38

39 Theoretical Analysis intro. to theory Markov chain problem dependency RTA analysis tools Fitness Level Method Noteonthat elitist never the best solu comparison with classics (i.e. on real-world situationslose summary on parameters sets, intuitively, form stairs, for which an u [Wegener, 02] a population is treated as the takenbest for getting offin every and a lower solution the stair, population Note that elitist (i.e. nb partition the(i.e. solution space in Lemma 2. S S m = Noteselect that elitist never loseformally the bestdescribed solution) EAs select solutions wit sets, intuitively, form est solution)note EAs solutions with better fitness. The level that elitist (i.e. never lose the best solution) EAs select be into subspaces solution... 2an Lemma (Fitness Level Method [39]) sets, intuitively, form stairs, for which upper bound can be derived taken for off ch an upper bound can be derived by summing upupper the time space sets, intuitively, form stairs, for bound can begetting derived byev su S2which an Sof For EA process on a problem f, get letol getting taken for gettingoffoff every stair, and aaelitist lower bound is the minimum formally described in lower boundtaken is thefor minimum time off stair. This getting every stair, and a an lower bound isisthe minimum timetime of with increasing fitness formallydescribed describedinin Lemma formally Lemma t = x) for all x 2 Si, and ui Then calculate: P ( t+ 2 [m j=i Lemma 2 (Fitness Leve process is at most Lemma 2 (Fitness Level Method [39]) Lemma 2 (Fitness Level Method [39]) For an elitist EA proces ). initialization probability of being in each subspace 0 (Si ) m X X -partition, m For an elitist EA process on a problem f, let S,..., S be < -partition, For an elitist EA process on a problem f, let S,..., S be a < let v le m m fa for m f, let S2.,..., Sm be <f -partition, let vi P ( t+ 2 [j=i+ bounds ofaprogress probability 0S (Sj i ) t = x), f all x 2 Si,ia for i : all x 2 S, and u x = 2x)Sfor P ( 2 [m m S = x) forvjall x 2 S. Then, j Sjt t j=i t t = x) for all x 2 isi, andiui Pt+ ( t+ i m 2j=i+ [j=i+ = x) for all x i2 Si. Th 2 [m j=i+ Sj t = x) for all x 2 Si. Then, the DCFHT of the EA process is at most process is atatmost process most and isand at least the ERT is thenisupper bounded by: lower bounded by: m X m 0 (Si ) m X X X X X 0 (Si ). i m 0 (S,, i )i ) 0 (S ui v j=i j vj i m i m j=i i m and is at least 39 and istoat least Optimization An Introduction Evolutionary

40 Example in OneMax partition 0 -bit -bit 2 -bits n -bits S 3... = S S S 2 S m initialization distribution: 0 (S i )= n i 2 n progress probability for : x 2 S i a lower bound: flipping one 0-bit but no -bits: n i ( n )(n n )n ERT: X appleiapplem 0 (S i ) mx j=i m X apple 0 (S 0 ) v j j= v j 2 O(n ln n) 40

41 Drift Analysis [Hajek, 82][Sasaki & Hajek, 88][He & Yao, 0][He & Yao, 04] x Then calculate: V(x) optimal solutions S. initialization probability of solutions 0 (x) 2. bounds of progress distance for every step: distance function V measuring distance of a solution to optimal solutions. V(x * )=0 P c l apple E[V ( t ) V ( t+ ) t ] c u E[V ( t ) V ( t+ ) t ] the ERT is then upper bounded by: P x2x 0(x)V (x)/c l and lower bounded by: P x2x 0(x)V (x)/c u most simplified version 4

42 Example in LeadingOnes LeadingOnes Problem: fitness: The drift: f(x) = nx iy i= j= arg max x2{0,} n x i nx i= j= Distance function: V (x) =n E[V ( t ) V ( t+ ) t ]= iy x i f(x) I(V ( t ) >V( t+ ))E[V ( t ) V ( t+ ) t ]+ I(V ( t ) <V( t+ ))E[V ( t ) V ( t+ ) t ]+ I(V ( t )=V ( t+ ))E[V ( t ) V ( t+ ) t ] Only need to care the expected progress: keep flip count the number of leading -bits f(0) = 2 distance of optimal solutions is zero zero zero probability of making progress >= probability of increasing at least one leading -bit E[V ( t ) V ( t+ ) t ] n ( n )i n ( n )n en 42

43 Example in LeadingOnes LeadingOnes Problem: fitness: f(x) = nx iy i= j= arg max x2{0,} n x i nx Distance function: V (x) =n iy i= j= x i f(x) count the number of leading -bits f(0) = 2 distance of optimal solutions is zero E[V ( t ) V ( t+ ) t ] n ( n )i n ( n )n en ERT is then upper bounded as X x2x 0 (x)v (x) en apple V ((00...0)) en = n en 2 O(n 2 ) 43

44 Additional Remarks There are more analytical tools than what we ve talked today. Different analytical tools may suit different algorithms/ problems (at least from an intuitive perspective). Theoretical analysis of EAs have been extended to more complicated problems (e.g., NP-hard problems such as Traveling Salesman, Minimum Vertex Cover, etc). In addition to deepening our understanding, theoretical analysis may also help configuring EAs for a specific problem. 44

45 End of Lecture 45

An Analysis on Recombination in Multi-Objective Evolutionary Optimization

An Analysis on Recombination in Multi-Objective Evolutionary Optimization An Analysis on Recombination in Multi-Objective Evolutionary Optimization Chao Qian, Yang Yu, Zhi-Hua Zhou National Key Laboratory for Novel Software Technology Nanjing University, Nanjing 20023, China

More information

Switch Analysis for Running Time Analysis of Evolutionary Algorithms

Switch Analysis for Running Time Analysis of Evolutionary Algorithms IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. XX, NO. X, 204 Switch Analysis for Running Time Analysis of Evolutionary Algorithms Yang Yu, Member, IEEE, Chao Qian, Zhi-Hua Zhou, Fellow, IEEE Abstract

More information

A Lower Bound Analysis of Population-based Evolutionary Algorithms for Pseudo-Boolean Functions

A Lower Bound Analysis of Population-based Evolutionary Algorithms for Pseudo-Boolean Functions A Lower Bound Analysis of Population-based Evolutionary Algorithms for Pseudo-Boolean Functions Chao Qian,2, Yang Yu 2, and Zhi-Hua Zhou 2 UBRI, School of Computer Science and Technology, University of

More information

A Comparison of GAs Penalizing Infeasible Solutions and Repairing Infeasible Solutions on the 0-1 Knapsack Problem

A Comparison of GAs Penalizing Infeasible Solutions and Repairing Infeasible Solutions on the 0-1 Knapsack Problem A Comparison of GAs Penalizing Infeasible Solutions and Repairing Infeasible Solutions on the 0-1 Knapsack Problem Jun He 1, Yuren Zhou 2, and Xin Yao 3 1 J. He is with the Department of Computer Science,

More information

Algorithms. Grad Refresher Course 2011 University of British Columbia. Ron Maharik

Algorithms. Grad Refresher Course 2011 University of British Columbia. Ron Maharik Algorithms Grad Refresher Course 2011 University of British Columbia Ron Maharik maharik@cs.ubc.ca About this talk For those incoming grad students who Do not have a CS background, or Have a CS background

More information

Evolutionary Computation Theory. Jun He School of Computer Science University of Birmingham Web: jxh

Evolutionary Computation Theory. Jun He School of Computer Science University of Birmingham Web:   jxh Evolutionary Computation Theory Jun He School of Computer Science University of Birmingham Web: www.cs.bham.ac.uk/ jxh Outline Motivation History Schema Theorem Convergence and Convergence Rate Computational

More information

How to Treat Evolutionary Algorithms as Ordinary Randomized Algorithms

How to Treat Evolutionary Algorithms as Ordinary Randomized Algorithms How to Treat Evolutionary Algorithms as Ordinary Randomized Algorithms Technical University of Denmark (Workshop Computational Approaches to Evolution, March 19, 2014) 1/22 Context Running time/complexity

More information

A New Approach to Estimating the Expected First Hitting Time of Evolutionary Algorithms

A New Approach to Estimating the Expected First Hitting Time of Evolutionary Algorithms A New Approach to Estimating the Expected First Hitting Time of Evolutionary Algorithms Yang Yu and Zhi-Hua Zhou National Laboratory for Novel Software Technology Nanjing University, Nanjing 20093, China

More information

DRAFT -- DRAFT -- DRAFT -- DRAFT -- DRAFT --

DRAFT -- DRAFT -- DRAFT -- DRAFT -- DRAFT -- Towards an Analytic Framework for Analysing the Computation Time of Evolutionary Algorithms Jun He and Xin Yao Abstract In spite of many applications of evolutionary algorithms in optimisation, theoretical

More information

On the Usefulness of Infeasible Solutions in Evolutionary Search: A Theoretical Study

On the Usefulness of Infeasible Solutions in Evolutionary Search: A Theoretical Study On the Usefulness of Infeasible Solutions in Evolutionary Search: A Theoretical Study Yang Yu, and Zhi-Hua Zhou, Senior Member, IEEE National Key Laboratory for Novel Software Technology Nanjing University,

More information

A Gentle Introduction to the Time Complexity Analysis of Evolutionary Algorithms

A Gentle Introduction to the Time Complexity Analysis of Evolutionary Algorithms A Gentle Introduction to the Time Complexity Analysis of Evolutionary Algorithms Pietro S. Oliveto Department of Computer Science, University of Sheffield, UK Symposium Series in Computational Intelligence

More information

Genetic Algorithms: Basic Principles and Applications

Genetic Algorithms: Basic Principles and Applications Genetic Algorithms: Basic Principles and Applications C. A. MURTHY MACHINE INTELLIGENCE UNIT INDIAN STATISTICAL INSTITUTE 203, B.T.ROAD KOLKATA-700108 e-mail: murthy@isical.ac.in Genetic algorithms (GAs)

More information

More on NP and Reductions

More on NP and Reductions Indian Institute of Information Technology Design and Manufacturing, Kancheepuram Chennai 600 127, India An Autonomous Institute under MHRD, Govt of India http://www.iiitdm.ac.in COM 501 Advanced Data

More information

P, NP, NP-Complete. Ruth Anderson

P, NP, NP-Complete. Ruth Anderson P, NP, NP-Complete Ruth Anderson A Few Problems: Euler Circuits Hamiltonian Circuits Intractability: P and NP NP-Complete What now? Today s Agenda 2 Try it! Which of these can you draw (trace all edges)

More information

On the Effectiveness of Sampling for Evolutionary Optimization in Noisy Environments

On the Effectiveness of Sampling for Evolutionary Optimization in Noisy Environments On the Effectiveness of Sampling for Evolutionary Optimization in Noisy Environments Chao Qian 1, Yang Yu 1, Yaochu Jin 2, and Zhi-Hua Zhou 1 1 National Key Laboratory for Novel Software Technology, Nanjing

More information

Average Drift Analysis and Population Scalability

Average Drift Analysis and Population Scalability Average Drift Analysis and Population Scalability Jun He and Xin Yao Abstract This paper aims to study how the population size affects the computation time of evolutionary algorithms in a rigorous way.

More information

Efficient Non-domination Level Update Method for Steady-State Evolutionary Multi-objective. optimization

Efficient Non-domination Level Update Method for Steady-State Evolutionary Multi-objective. optimization Efficient Non-domination Level Update Method for Steady-State Evolutionary Multi-objective Optimization Ke Li, Kalyanmoy Deb, Fellow, IEEE, Qingfu Zhang, Senior Member, IEEE, and Qiang Zhang COIN Report

More information

Evolutionary Algorithms How to Cope With Plateaus of Constant Fitness and When to Reject Strings of The Same Fitness

Evolutionary Algorithms How to Cope With Plateaus of Constant Fitness and When to Reject Strings of The Same Fitness Evolutionary Algorithms How to Cope With Plateaus of Constant Fitness and When to Reject Strings of The Same Fitness Thomas Jansen and Ingo Wegener FB Informatik, LS 2, Univ. Dortmund, 44221 Dortmund,

More information

Computer Science 385 Analysis of Algorithms Siena College Spring Topic Notes: Limitations of Algorithms

Computer Science 385 Analysis of Algorithms Siena College Spring Topic Notes: Limitations of Algorithms Computer Science 385 Analysis of Algorithms Siena College Spring 2011 Topic Notes: Limitations of Algorithms We conclude with a discussion of the limitations of the power of algorithms. That is, what kinds

More information

20.1 2SAT. CS125 Lecture 20 Fall 2016

20.1 2SAT. CS125 Lecture 20 Fall 2016 CS125 Lecture 20 Fall 2016 20.1 2SAT We show yet another possible way to solve the 2SAT problem. Recall that the input to 2SAT is a logical expression that is the conunction (AND) of a set of clauses,

More information

Metaheuristics and Local Search

Metaheuristics and Local Search Metaheuristics and Local Search 8000 Discrete optimization problems Variables x 1,..., x n. Variable domains D 1,..., D n, with D j Z. Constraints C 1,..., C m, with C i D 1 D n. Objective function f :

More information

Geometric Semantic Genetic Programming (GSGP): theory-laden design of semantic mutation operators

Geometric Semantic Genetic Programming (GSGP): theory-laden design of semantic mutation operators Geometric Semantic Genetic Programming (GSGP): theory-laden design of semantic mutation operators Andrea Mambrini 1 University of Birmingham, Birmingham UK 6th June 2013 1 / 33 Andrea Mambrini GSGP: theory-laden

More information

NP-Completeness. Until now we have been designing algorithms for specific problems

NP-Completeness. Until now we have been designing algorithms for specific problems NP-Completeness 1 Introduction Until now we have been designing algorithms for specific problems We have seen running times O(log n), O(n), O(n log n), O(n 2 ), O(n 3 )... We have also discussed lower

More information

When Is an Estimation of Distribution Algorithm Better than an Evolutionary Algorithm?

When Is an Estimation of Distribution Algorithm Better than an Evolutionary Algorithm? When Is an Estimation of Distribution Algorithm Better than an Evolutionary Algorithm? Tianshi Chen, Per Kristian Lehre, Ke Tang, and Xin Yao Abstract Despite the wide-spread popularity of estimation of

More information

Running time analysis of a multi-objective evolutionary algorithm on a simple discrete optimization problem

Running time analysis of a multi-objective evolutionary algorithm on a simple discrete optimization problem Research Collection Working Paper Running time analysis of a multi-objective evolutionary algorithm on a simple discrete optimization problem Author(s): Laumanns, Marco; Thiele, Lothar; Zitzler, Eckart;

More information

Geometric Semantic Genetic Programming (GSGP): theory-laden design of variation operators

Geometric Semantic Genetic Programming (GSGP): theory-laden design of variation operators Geometric Semantic Genetic Programming (GSGP): theory-laden design of variation operators Andrea Mambrini University of Birmingham, UK NICaiA Exchange Programme LaMDA group, Nanjing University, China 7th

More information

Metaheuristics and Local Search. Discrete optimization problems. Solution approaches

Metaheuristics and Local Search. Discrete optimization problems. Solution approaches Discrete Mathematics for Bioinformatics WS 07/08, G. W. Klau, 31. Januar 2008, 11:55 1 Metaheuristics and Local Search Discrete optimization problems Variables x 1,...,x n. Variable domains D 1,...,D n,

More information

Introduction to Optimization

Introduction to Optimization Introduction to Optimization Blackbox Optimization Marc Toussaint U Stuttgart Blackbox Optimization The term is not really well defined I use it to express that only f(x) can be evaluated f(x) or 2 f(x)

More information

Limitations of Algorithm Power

Limitations of Algorithm Power Limitations of Algorithm Power Objectives We now move into the third and final major theme for this course. 1. Tools for analyzing algorithms. 2. Design strategies for designing algorithms. 3. Identifying

More information

Algorithm Design Strategies V

Algorithm Design Strategies V Algorithm Design Strategies V Joaquim Madeira Version 0.0 October 2016 U. Aveiro, October 2016 1 Overview The 0-1 Knapsack Problem Revisited The Fractional Knapsack Problem Greedy Algorithms Example Coin

More information

Computational complexity theory

Computational complexity theory Computational complexity theory Introduction to computational complexity theory Complexity (computability) theory deals with two aspects: Algorithm s complexity. Problem s complexity. References S. Cook,

More information

Running Time Analysis of Multi-objective Evolutionary Algorithms on a Simple Discrete Optimization Problem

Running Time Analysis of Multi-objective Evolutionary Algorithms on a Simple Discrete Optimization Problem Running Time Analysis of Multi-objective Evolutionary Algorithms on a Simple Discrete Optimization Problem Marco Laumanns 1, Lothar Thiele 1, Eckart Zitzler 1,EmoWelzl 2,and Kalyanmoy Deb 3 1 ETH Zürich,

More information

CS 781 Lecture 9 March 10, 2011 Topics: Local Search and Optimization Metropolis Algorithm Greedy Optimization Hopfield Networks Max Cut Problem Nash

CS 781 Lecture 9 March 10, 2011 Topics: Local Search and Optimization Metropolis Algorithm Greedy Optimization Hopfield Networks Max Cut Problem Nash CS 781 Lecture 9 March 10, 2011 Topics: Local Search and Optimization Metropolis Algorithm Greedy Optimization Hopfield Networks Max Cut Problem Nash Equilibrium Price of Stability Coping With NP-Hardness

More information

Data Structures in Java

Data Structures in Java Data Structures in Java Lecture 21: Introduction to NP-Completeness 12/9/2015 Daniel Bauer Algorithms and Problem Solving Purpose of algorithms: find solutions to problems. Data Structures provide ways

More information

REIHE COMPUTATIONAL INTELLIGENCE COLLABORATIVE RESEARCH CENTER 531

REIHE COMPUTATIONAL INTELLIGENCE COLLABORATIVE RESEARCH CENTER 531 U N I V E R S I T Y OF D O R T M U N D REIHE COMPUTATIONAL INTELLIGENCE COLLABORATIVE RESEARCH CENTER 53 Design and Management of Complex Technical Processes and Systems by means of Computational Intelligence

More information

Black Box Search By Unbiased Variation

Black Box Search By Unbiased Variation Black Box Search By Unbiased Variation Per Kristian Lehre and Carsten Witt CERCIA, University of Birmingham, UK DTU Informatics, Copenhagen, Denmark ThRaSH - March 24th 2010 State of the Art in Runtime

More information

NP-Completeness. Andreas Klappenecker. [based on slides by Prof. Welch]

NP-Completeness. Andreas Klappenecker. [based on slides by Prof. Welch] NP-Completeness Andreas Klappenecker [based on slides by Prof. Welch] 1 Prelude: Informal Discussion (Incidentally, we will never get very formal in this course) 2 Polynomial Time Algorithms Most of the

More information

When to Use Bit-Wise Neutrality

When to Use Bit-Wise Neutrality When to Use Bit-Wise Neutrality Tobias Friedrich Department 1: Algorithms and Complexity Max-Planck-Institut für Informatik Saarbrücken, Germany Frank Neumann Department 1: Algorithms and Complexity Max-Planck-Institut

More information

NP-Completeness. CptS 223 Advanced Data Structures. Larry Holder School of Electrical Engineering and Computer Science Washington State University

NP-Completeness. CptS 223 Advanced Data Structures. Larry Holder School of Electrical Engineering and Computer Science Washington State University NP-Completeness CptS 223 Advanced Data Structures Larry Holder School of Electrical Engineering and Computer Science Washington State University 1 Hard Graph Problems Hard means no known solutions with

More information

5. Simulated Annealing 5.1 Basic Concepts. Fall 2010 Instructor: Dr. Masoud Yaghini

5. Simulated Annealing 5.1 Basic Concepts. Fall 2010 Instructor: Dr. Masoud Yaghini 5. Simulated Annealing 5.1 Basic Concepts Fall 2010 Instructor: Dr. Masoud Yaghini Outline Introduction Real Annealing and Simulated Annealing Metropolis Algorithm Template of SA A Simple Example References

More information

COMS W4995 Introduction to Cryptography October 12, Lecture 12: RSA, and a summary of One Way Function Candidates.

COMS W4995 Introduction to Cryptography October 12, Lecture 12: RSA, and a summary of One Way Function Candidates. COMS W4995 Introduction to Cryptography October 12, 2005 Lecture 12: RSA, and a summary of One Way Function Candidates. Lecturer: Tal Malkin Scribes: Justin Cranshaw and Mike Verbalis 1 Introduction In

More information

Spin Glas Dynamics and Stochastic Optimization Schemes. Karl Heinz Hoffmann TU Chemnitz

Spin Glas Dynamics and Stochastic Optimization Schemes. Karl Heinz Hoffmann TU Chemnitz Spin Glas Dynamics and Stochastic Optimization Schemes Karl Heinz Hoffmann TU Chemnitz 1 Spin Glasses spin glass e.g. AuFe AuMn CuMn nobel metal (no spin) transition metal (spin) 0.1-10 at% ferromagnetic

More information

Mathematics for Decision Making: An Introduction. Lecture 8

Mathematics for Decision Making: An Introduction. Lecture 8 Mathematics for Decision Making: An Introduction Lecture 8 Matthias Köppe UC Davis, Mathematics January 29, 2009 8 1 Shortest Paths and Feasible Potentials Feasible Potentials Suppose for all v V, there

More information

Advanced Optimization

Advanced Optimization Advanced Optimization Lecture 3: 1: Randomized Algorithms for for Continuous Discrete Problems Problems November 22, 2016 Master AIC Université Paris-Saclay, Orsay, France Anne Auger INRIA Saclay Ile-de-France

More information

Data Warehousing & Data Mining

Data Warehousing & Data Mining 13. Meta-Algorithms for Classification Data Warehousing & Data Mining Wolf-Tilo Balke Silviu Homoceanu Institut für Informationssysteme Technische Universität Braunschweig http://www.ifis.cs.tu-bs.de 13.

More information

Definition (The carefully thought-out calculus version based on limits).

Definition (The carefully thought-out calculus version based on limits). 4.1. Continuity and Graphs Definition 4.1.1 (Intuitive idea used in algebra based on graphing). A function, f, is continuous on the interval (a, b) if the graph of y = f(x) can be drawn over the interval

More information

CSC 4510 Machine Learning

CSC 4510 Machine Learning 10: Gene(c Algorithms CSC 4510 Machine Learning Dr. Mary Angela Papalaskari Department of CompuBng Sciences Villanova University Course website: www.csc.villanova.edu/~map/4510/ Slides of this presenta(on

More information

Lecture 22. Introduction to Genetic Algorithms

Lecture 22. Introduction to Genetic Algorithms Lecture 22 Introduction to Genetic Algorithms Thursday 14 November 2002 William H. Hsu, KSU http://www.kddresearch.org http://www.cis.ksu.edu/~bhsu Readings: Sections 9.1-9.4, Mitchell Chapter 1, Sections

More information

Lecture 9 Evolutionary Computation: Genetic algorithms

Lecture 9 Evolutionary Computation: Genetic algorithms Lecture 9 Evolutionary Computation: Genetic algorithms Introduction, or can evolution be intelligent? Simulation of natural evolution Genetic algorithms Case study: maintenance scheduling with genetic

More information

Complexity (Pre Lecture)

Complexity (Pre Lecture) Complexity (Pre Lecture) Dr. Neil T. Dantam CSCI-561, Colorado School of Mines Fall 2018 Dantam (Mines CSCI-561) Complexity (Pre Lecture) Fall 2018 1 / 70 Why? What can we always compute efficiently? What

More information

Pure Strategy or Mixed Strategy?

Pure Strategy or Mixed Strategy? Pure Strategy or Mixed Strategy? Jun He, Feidun He, Hongbin Dong arxiv:257v4 [csne] 4 Apr 204 Abstract Mixed strategy evolutionary algorithms EAs) aim at integrating several mutation operators into a single

More information

Zebo Peng Embedded Systems Laboratory IDA, Linköping University

Zebo Peng Embedded Systems Laboratory IDA, Linköping University TDTS 01 Lecture 8 Optimization Heuristics for Synthesis Zebo Peng Embedded Systems Laboratory IDA, Linköping University Lecture 8 Optimization problems Heuristic techniques Simulated annealing Genetic

More information

REIHE COMPUTATIONAL INTELLIGENCE COLLABORATIVE RESEARCH CENTER 531

REIHE COMPUTATIONAL INTELLIGENCE COLLABORATIVE RESEARCH CENTER 531 U N I V E R S I T Y OF D O R T M U N D REIHE COMPUTATIONAL INTELLIGENCE COLLABORATIVE RESEARCH CENTER 531 Design and Management of Complex Technical Processes and Systems by means of Computational Intelligence

More information

Lecture notes for Analysis of Algorithms : Markov decision processes

Lecture notes for Analysis of Algorithms : Markov decision processes Lecture notes for Analysis of Algorithms : Markov decision processes Lecturer: Thomas Dueholm Hansen June 6, 013 Abstract We give an introduction to infinite-horizon Markov decision processes (MDPs) with

More information

Computational complexity theory

Computational complexity theory Computational complexity theory Introduction to computational complexity theory Complexity (computability) theory deals with two aspects: Algorithm s complexity. Problem s complexity. References S. Cook,

More information

Turing Machines and Time Complexity

Turing Machines and Time Complexity Turing Machines and Time Complexity Turing Machines Turing Machines (Infinitely long) Tape of 1 s and 0 s Turing Machines (Infinitely long) Tape of 1 s and 0 s Able to read and write the tape, and move

More information

Introduction to Algorithms / Algorithms I Lecturer: Michael Dinitz Topic: Intro to Learning Theory Date: 12/8/16

Introduction to Algorithms / Algorithms I Lecturer: Michael Dinitz Topic: Intro to Learning Theory Date: 12/8/16 600.463 Introduction to Algorithms / Algorithms I Lecturer: Michael Dinitz Topic: Intro to Learning Theory Date: 12/8/16 25.1 Introduction Today we re going to talk about machine learning, but from an

More information

On the Impact of Objective Function Transformations on Evolutionary and Black-Box Algorithms

On the Impact of Objective Function Transformations on Evolutionary and Black-Box Algorithms On the Impact of Objective Function Transformations on Evolutionary and Black-Box Algorithms [Extended Abstract] Tobias Storch Department of Computer Science 2, University of Dortmund, 44221 Dortmund,

More information

Distributed Optimization. Song Chong EE, KAIST

Distributed Optimization. Song Chong EE, KAIST Distributed Optimization Song Chong EE, KAIST songchong@kaist.edu Dynamic Programming for Path Planning A path-planning problem consists of a weighted directed graph with a set of n nodes N, directed links

More information

Hertz, Krogh, Palmer: Introduction to the Theory of Neural Computation. Addison-Wesley Publishing Company (1991). (v ji (1 x i ) + (1 v ji )x i )

Hertz, Krogh, Palmer: Introduction to the Theory of Neural Computation. Addison-Wesley Publishing Company (1991). (v ji (1 x i ) + (1 v ji )x i ) Symmetric Networks Hertz, Krogh, Palmer: Introduction to the Theory of Neural Computation. Addison-Wesley Publishing Company (1991). How can we model an associative memory? Let M = {v 1,..., v m } be a

More information

Lecture notes for quantum semidefinite programming (SDP) solvers

Lecture notes for quantum semidefinite programming (SDP) solvers CMSC 657, Intro to Quantum Information Processing Lecture on November 15 and 0, 018 Fall 018, University of Maryland Prepared by Tongyang Li, Xiaodi Wu Lecture notes for quantum semidefinite programming

More information

Computational hardness. Feb 2 abhi shelat

Computational hardness. Feb 2 abhi shelat L4 6501 Computational hardness Feb 2 abhi shelat Eve Alice Bob Eve Alice Bob k Gen k Eve Alice Bob c=enck(mi) k Gen k Eve c Alice Bob c=enck(mi) k Gen k Eve c Alice c=enck(mi) Bob m=deck(c) k Gen k Eve

More information

Non-Convex Optimization. CS6787 Lecture 7 Fall 2017

Non-Convex Optimization. CS6787 Lecture 7 Fall 2017 Non-Convex Optimization CS6787 Lecture 7 Fall 2017 First some words about grading I sent out a bunch of grades on the course management system Everyone should have all their grades in Not including paper

More information

Theory of Computer Science

Theory of Computer Science Theory of Computer Science E1. Complexity Theory: Motivation and Introduction Malte Helmert University of Basel May 18, 2016 Overview: Course contents of this course: logic How can knowledge be represented?

More information

Empirical Risk Minimization

Empirical Risk Minimization Empirical Risk Minimization Fabrice Rossi SAMM Université Paris 1 Panthéon Sorbonne 2018 Outline Introduction PAC learning ERM in practice 2 General setting Data X the input space and Y the output space

More information

P P P NP-Hard: L is NP-hard if for all L NP, L L. Thus, if we could solve L in polynomial. Cook's Theorem and Reductions

P P P NP-Hard: L is NP-hard if for all L NP, L L. Thus, if we could solve L in polynomial. Cook's Theorem and Reductions Summary of the previous lecture Recall that we mentioned the following topics: P: is the set of decision problems (or languages) that are solvable in polynomial time. NP: is the set of decision problems

More information

PROBLEM SOLVING AND SEARCH IN ARTIFICIAL INTELLIGENCE

PROBLEM SOLVING AND SEARCH IN ARTIFICIAL INTELLIGENCE Artificial Intelligence, Computational Logic PROBLEM SOLVING AND SEARCH IN ARTIFICIAL INTELLIGENCE Lecture 4 Metaheuristic Algorithms Sarah Gaggl Dresden, 5th May 2017 Agenda 1 Introduction 2 Constraint

More information

Runtime Analysis of Evolutionary Algorithms for the Knapsack Problem with Favorably Correlated Weights

Runtime Analysis of Evolutionary Algorithms for the Knapsack Problem with Favorably Correlated Weights Runtime Analysis of Evolutionary Algorithms for the Knapsack Problem with Favorably Correlated Weights Frank Neumann 1 and Andrew M. Sutton 2 1 Optimisation and Logistics, School of Computer Science, The

More information

Plateaus Can Be Harder in Multi-Objective Optimization

Plateaus Can Be Harder in Multi-Objective Optimization Plateaus Can Be Harder in Multi-Objective Optimization Tobias Friedrich and Nils Hebbinghaus and Frank Neumann Max-Planck-Institut für Informatik, Campus E1 4, 66123 Saarbrücken, Germany Abstract In recent

More information

Theory of Computer Science. Theory of Computer Science. E1.1 Motivation. E1.2 How to Measure Runtime? E1.3 Decision Problems. E1.

Theory of Computer Science. Theory of Computer Science. E1.1 Motivation. E1.2 How to Measure Runtime? E1.3 Decision Problems. E1. Theory of Computer Science May 18, 2016 E1. Complexity Theory: Motivation and Introduction Theory of Computer Science E1. Complexity Theory: Motivation and Introduction Malte Helmert University of Basel

More information

Runtime Analysis of a Binary Particle Swarm Optimizer

Runtime Analysis of a Binary Particle Swarm Optimizer Runtime Analysis of a Binary Particle Swarm Optimizer Dirk Sudholt Fakultät für Informatik, LS 2 Technische Universität Dortmund Dortmund, Germany Carsten Witt Fakultät für Informatik, LS 2 Technische

More information

Lecture 4: Simulated Annealing. An Introduction to Meta-Heuristics, Produced by Qiangfu Zhao (Since 2012), All rights reserved

Lecture 4: Simulated Annealing. An Introduction to Meta-Heuristics, Produced by Qiangfu Zhao (Since 2012), All rights reserved Lecture 4: Simulated Annealing Lec04/1 Neighborhood based search Step 1: s=s0; Step 2: Stop if terminating condition satisfied; Step 3: Generate N(s); Step 4: s =FindBetterSolution(N(s)); Step 5: s=s ;

More information

R ij = 2. Using all of these facts together, you can solve problem number 9.

R ij = 2. Using all of these facts together, you can solve problem number 9. Help for Homework Problem #9 Let G(V,E) be any undirected graph We want to calculate the travel time across the graph. Think of each edge as one resistor of 1 Ohm. Say we have two nodes: i and j Let the

More information

Complexity Theory VU , SS The Polynomial Hierarchy. Reinhard Pichler

Complexity Theory VU , SS The Polynomial Hierarchy. Reinhard Pichler Complexity Theory Complexity Theory VU 181.142, SS 2018 6. The Polynomial Hierarchy Reinhard Pichler Institut für Informationssysteme Arbeitsbereich DBAI Technische Universität Wien 15 May, 2018 Reinhard

More information

Outline. Complexity Theory EXACT TSP. The Class DP. Definition. Problem EXACT TSP. Complexity of EXACT TSP. Proposition VU 181.

Outline. Complexity Theory EXACT TSP. The Class DP. Definition. Problem EXACT TSP. Complexity of EXACT TSP. Proposition VU 181. Complexity Theory Complexity Theory Outline Complexity Theory VU 181.142, SS 2018 6. The Polynomial Hierarchy Reinhard Pichler Institut für Informationssysteme Arbeitsbereich DBAI Technische Universität

More information

6. APPLICATION TO THE TRAVELING SALESMAN PROBLEM

6. APPLICATION TO THE TRAVELING SALESMAN PROBLEM 6. Application to the Traveling Salesman Problem 92 6. APPLICATION TO THE TRAVELING SALESMAN PROBLEM The properties that have the most significant influence on the maps constructed by Kohonen s algorithm

More information

Planning in Markov Decision Processes

Planning in Markov Decision Processes Carnegie Mellon School of Computer Science Deep Reinforcement Learning and Control Planning in Markov Decision Processes Lecture 3, CMU 10703 Katerina Fragkiadaki Markov Decision Process (MDP) A Markov

More information

Runtime Analyses for Using Fairness in Evolutionary Multi-Objective Optimization

Runtime Analyses for Using Fairness in Evolutionary Multi-Objective Optimization Runtime Analyses for Using Fairness in Evolutionary Multi-Objective Optimization Tobias Friedrich 1, Christian Horoba 2, and Frank Neumann 1 1 Max-Planck-Institut für Informatik, Saarbrücken, Germany 2

More information

Computational Complexity

Computational Complexity p. 1/24 Computational Complexity The most sharp distinction in the theory of computation is between computable and noncomputable functions; that is, between possible and impossible. From the example of

More information

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008 Gaussian processes Chuong B Do (updated by Honglak Lee) November 22, 2008 Many of the classical machine learning algorithms that we talked about during the first half of this course fit the following pattern:

More information

Methods for finding optimal configurations

Methods for finding optimal configurations CS 1571 Introduction to AI Lecture 9 Methods for finding optimal configurations Milos Hauskrecht milos@cs.pitt.edu 5329 Sennott Square Search for the optimal configuration Optimal configuration search:

More information

Queens College, CUNY, Department of Computer Science Numerical Methods CSCI 361 / 761 Spring 2018 Instructor: Dr. Sateesh Mane.

Queens College, CUNY, Department of Computer Science Numerical Methods CSCI 361 / 761 Spring 2018 Instructor: Dr. Sateesh Mane. Queens College, CUNY, Department of Computer Science Numerical Methods CSCI 361 / 761 Spring 2018 Instructor: Dr. Sateesh Mane c Sateesh R. Mane 2018 3 Lecture 3 3.1 General remarks March 4, 2018 This

More information

In complexity theory, algorithms and problems are classified by the growth order of computation time as a function of instance size.

In complexity theory, algorithms and problems are classified by the growth order of computation time as a function of instance size. 10 2.2. CLASSES OF COMPUTATIONAL COMPLEXITY An optimization problem is defined as a class of similar problems with different input parameters. Each individual case with fixed parameter values is called

More information

NP-Completeness I. Lecture Overview Introduction: Reduction and Expressiveness

NP-Completeness I. Lecture Overview Introduction: Reduction and Expressiveness Lecture 19 NP-Completeness I 19.1 Overview In the past few lectures we have looked at increasingly more expressive problems that we were able to solve using efficient algorithms. In this lecture we introduce

More information

CHI SQUARE ANALYSIS 8/18/2011 HYPOTHESIS TESTS SO FAR PARAMETRIC VS. NON-PARAMETRIC

CHI SQUARE ANALYSIS 8/18/2011 HYPOTHESIS TESTS SO FAR PARAMETRIC VS. NON-PARAMETRIC CHI SQUARE ANALYSIS I N T R O D U C T I O N T O N O N - P A R A M E T R I C A N A L Y S E S HYPOTHESIS TESTS SO FAR We ve discussed One-sample t-test Dependent Sample t-tests Independent Samples t-tests

More information

Restarting a Genetic Algorithm for Set Cover Problem Using Schnabel Census

Restarting a Genetic Algorithm for Set Cover Problem Using Schnabel Census Restarting a Genetic Algorithm for Set Cover Problem Using Schnabel Census Anton V. Eremeev 1,2 1 Dostoevsky Omsk State University, Omsk, Russia 2 The Institute of Scientific Information for Social Sciences

More information

6.867 Machine Learning

6.867 Machine Learning 6.867 Machine Learning Problem set 1 Solutions Thursday, September 19 What and how to turn in? Turn in short written answers to the questions explicitly stated, and when requested to explain or prove.

More information

The Fitness Level Method with Tail Bounds

The Fitness Level Method with Tail Bounds The Fitness Level Method with Tail Bounds Carsten Witt DTU Compute Technical University of Denmark 2800 Kgs. Lyngby Denmark arxiv:307.4274v [cs.ne] 6 Jul 203 July 7, 203 Abstract The fitness-level method,

More information

Chapter 8: Introduction to Evolutionary Computation

Chapter 8: Introduction to Evolutionary Computation Computational Intelligence: Second Edition Contents Some Theories about Evolution Evolution is an optimization process: the aim is to improve the ability of an organism to survive in dynamically changing

More information

Simulated Annealing Beats Metropolis in Combinatorial Optimization

Simulated Annealing Beats Metropolis in Combinatorial Optimization Electronic Colloquium on Computational Complexity, Report No. 89 (2004) Simulated Annealing Beats Metropolis in Combinatorial Optimization Ingo Wegener FB Informatik, LS2, Univ. Dortmund, Germany ingo.wegener@uni-dortmund.de

More information

Algorithms other than SGD. CS6787 Lecture 10 Fall 2017

Algorithms other than SGD. CS6787 Lecture 10 Fall 2017 Algorithms other than SGD CS6787 Lecture 10 Fall 2017 Machine learning is not just SGD Once a model is trained, we need to use it to classify new examples This inference task is not computed with SGD There

More information

COMS 4721: Machine Learning for Data Science Lecture 20, 4/11/2017

COMS 4721: Machine Learning for Data Science Lecture 20, 4/11/2017 COMS 4721: Machine Learning for Data Science Lecture 20, 4/11/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University SEQUENTIAL DATA So far, when thinking

More information

Runtime Analysis of Genetic Algorithms with Very High Selection Pressure

Runtime Analysis of Genetic Algorithms with Very High Selection Pressure Runtime Analysis of Genetic Algorithms with Very High Selection Pressure Anton V. Eremeev 1,2 1 Sobolev Institute of Mathematics, Omsk Branch, 13 Pevtsov str., 644099, Omsk, Russia 2 Omsk State University

More information

NP-Completeness. ch34 Hewett. Problem. Tractable Intractable Non-computable computationally infeasible super poly-time alg. sol. E.g.

NP-Completeness. ch34 Hewett. Problem. Tractable Intractable Non-computable computationally infeasible super poly-time alg. sol. E.g. NP-Completeness ch34 Hewett Problem Tractable Intractable Non-computable computationally infeasible super poly-time alg. sol. E.g., O(2 n ) computationally feasible poly-time alg. sol. E.g., O(n k ) No

More information

CS6901: review of Theory of Computation and Algorithms

CS6901: review of Theory of Computation and Algorithms CS6901: review of Theory of Computation and Algorithms Any mechanically (automatically) discretely computation of problem solving contains at least three components: - problem description - computational

More information

Markov chains. Randomness and Computation. Markov chains. Markov processes

Markov chains. Randomness and Computation. Markov chains. Markov processes Markov chains Randomness and Computation or, Randomized Algorithms Mary Cryan School of Informatics University of Edinburgh Definition (Definition 7) A discrete-time stochastic process on the state space

More information

Lecture 25: Cook s Theorem (1997) Steven Skiena. skiena

Lecture 25: Cook s Theorem (1997) Steven Skiena.   skiena Lecture 25: Cook s Theorem (1997) Steven Skiena Department of Computer Science State University of New York Stony Brook, NY 11794 4400 http://www.cs.sunysb.edu/ skiena Prove that Hamiltonian Path is NP

More information

Lecture 3. 1 Terminology. 2 Non-Deterministic Space Complexity. Notes on Complexity Theory: Fall 2005 Last updated: September, 2005.

Lecture 3. 1 Terminology. 2 Non-Deterministic Space Complexity. Notes on Complexity Theory: Fall 2005 Last updated: September, 2005. Notes on Complexity Theory: Fall 2005 Last updated: September, 2005 Jonathan Katz Lecture 3 1 Terminology For any complexity class C, we define the class coc as follows: coc def = { L L C }. One class

More information

/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: NP-Completeness I Date: 11/13/18

/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: NP-Completeness I Date: 11/13/18 601.433/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: NP-Completeness I Date: 11/13/18 20.1 Introduction Definition 20.1.1 We say that an algorithm runs in polynomial time if its running

More information

Theory of Computation

Theory of Computation Theory of Computation Dr. Sarmad Abbasi Dr. Sarmad Abbasi () Theory of Computation 1 / 33 Lecture 20: Overview Incompressible strings Minimal Length Descriptions Descriptive Complexity Dr. Sarmad Abbasi

More information