Running Time Analysis of a Multiobjective Evolutionary Algorithm on Simple and Hard Problems

Size: px
Start display at page:

Download "Running Time Analysis of a Multiobjective Evolutionary Algorithm on Simple and Hard Problems"

Transcription

1 112 Running Time Analysis of a Multiobjective Evolutionary Algorithm on Simple and Hard Problems Rajeev Kumar and Nilanjan Banerjee Department of Computer Science and Engineering Indian Institute of Technology Kharagpur Kharagpur, WB , India rkumar@cse.iitkgp.ernet.in, nilanb@cs.umass.edu Abstract. In this paper, we suggest a multiobjective evolutionary algorithm based on a restricted mating pool (REMO) with a separate archive for storing the remaining population. Such archive based algorithms have been used for solving real-world applications, however, no theoretical results are available. In this paper, we present a rigorous running time complexity analysis for the algorithm on two simple discrete pseudo boolean functions and on the multiobjective knapsack problem which is known to be NP-complete. We use two well known simple functions LOTZ (Leading Zeros: Trailing Ones) and a quadratic function. For the knapsack problem we formalize a (1 + ɛ)-approximation set under a constraint on the weights of the items. We then generalize the idea by eliminating the constraints based on a principle of partitioning the items into blocks and analyze REMO on it. We use a simple strategy based on partitioning of the decision space into fitness layers for the analysis. 1 Introduction Evolutionary algorithms are emerging as a powerful tool to solve NP-hard combinatorial optimization problems. EAs use a randomized search technique with a population of individuals. The genetic operators used by EAs do not apply, in general, any problem-specific knowledge, however, special genetic operators may be designed by incorporating domain knowledge to expedite the search for some applications. In the multiobjective scenario, EAs often find effectively a set of diverse and mutually competitive solutions. Some results for solving computationally hard problems [1, 2] using multiobjective EAs are available in the literature e.g., m-dimensional knapsack [3], minimum spanning tree [4, 5], partitioning of high-dimensional patterns spaces [6], code-book design [7], communication network topology design [8], and network design [9]. The EA operators like mutation and crossover imitate the process of natural evolution. The underlying principles of the operators are simple, but nevertheless, EAs exhibit complex behavior which is difficult to analyze theoretically. Hence, though there are numerous empirical reports on the application of EAs, work on their theoretical analysis is rare. However, besides empirical findings, theoretical analysis is essential to understand the performance and behavior of such heuristics. Some work, in this direction, has recently started, e.g., [10 12]. In case of single objective optimization, many results have been obtained on the analysis of evolutionary algorithms. Results on the time bounds of algorithms in the A.H. Wright et al. (Eds.): FOGA 2005, LNCS 3469, pp , c Springer-Verlag Berlin Heidelberg 2005

2 Running Time Analysis of a Multiobjective Evolutionary Algorithm 113 discrete search space [13] as well as continuous search space [14] are available. Some analysis on special functions using (1 + 1) EA has been done - linear functions [15], ONE-MAX function [16], unimodal function [13, 17], and pseudo-boolean quadratic functions [18]. Most of the work above analyzed evolutionary algorithms with mutation as the only genetic operator. However, a proof that crossover is essential is presented in [19]. The analysis of the multiobjective case, however, is more difficult than its single objective counterpart since it involves issues like the size of Pareto-set, diversity of the obtained solutions and convergence to the Pareto-front [20, 21]. Consequently, results on theoretical analysis of multiobjective evolutionary algorithms are few. Rudolph [22, 23] and Rudolph and Agapie [24] have studied multiobjective optimizers with respect to their limit behavior. Laumanns et al. pioneered in deriving sharp asymptotic bounds for two-objective toy functions [10, 11]. Recently, Giel [25] and Thierens [26] derived bounds for another simple function. Most of the work done earlier deals with analysis of simple problems. However, analysis of a multiobjective evolutionary algorithm on a simple variant of the 0-1 knapsack problem was started by Laumanns et al. [12]. They analyzed the expected running time of two multiobjective evolutionary algorithms Simple Evolutionary Multiobjective Optimizer (SEMO) and Fair Evolutionary Multiobjective Optimizer (FEMO) for a simple instance of the multiobjective 0-1 knapsack problem. The considered problem instance has two profit values per item and cannot be solved by one-bit mutations. In the analysis, the authors make use of two general upper bound techniques, the decision space partition method and the graph search method. The paper demonstrates how these methods, which have previously only been applied to algorithms with one-bit mutations, are equally applicable for mutation operators where each bit is flipped independently with a certain probability. However, the work takes care of only a very special instance of the knapsack and cannot be extended to any knapsack problem. None of the work involves analysis of real-world combinatorial optimization problems using multiobjective EAs. In this work, we continue such an analysis for the wellknown bi-objective 0 1 knapsack problem [27 30]. In the most general case, the Paretooptimal set for the knapsack can be exponentially large in the input size. Therefore, we first, formulate a (1 + ɛ)-approximate set for the 0-1 knapsack, followed by a rigorous analysis of the expected time to find the solution-set. We also carry out the expected running time analysis on two simple pseudo-boolean functions, namely the Leading Zeros: Trailing Ones (LOTZ) [10] and Quadratic Function (QF) [25]. We suggest a simple multiobjective optimizer based on an archiving strategy which is well adapted to work efficiently for problems where the Pareto-optimal points are Hamming neighbors (i.e., having a Hamming distance of 1). We call our algorithm Restricted Evolutionary Multiobjective Optimizer (REMO). The algorithm uses a special archive of two individuals which are selected based on a special fitness function which selects two individuals with the largest Hamming distance. Such a mechanism assures that the individuals selected for mutation are more likely to produce new individuals. However, this assumption holds for functions where the optimal set consists of individuals which are Hamming neighbors of each other and the distribution of the optimal points is fairly uniform.

3 114 Rajeev Kumar and Nilanjan Banerjee The rest of the paper is organized as follows. Section 2 describes the related work in the field of theoretical analysis of evolutionary algorithms. Section 3 includes a few definitions pertaining to multiobjective optimization. Section 4 describes our algorithm REMO. Section 5 presents the analysis of the REMO on the LOTZ and the Quadratic function. Section 6 formulates the linear functions, the knapsack problem and its (1+ɛ)- approximate set. The analysis of the algorithm on the knapsack problem is given in section 7. Finally, conclusions are drawn in section 8. 2 Related Work 2.1 Problems Analyzed The work of Beyer et al. [31] revolves around how long a particular algorithm takes to find the optimal solutions for a given class of functions. The motivation to start such an analysis was to improve the knowledge of the randomized search heuristics on a given class of functions. Rudolph [22, 23] and Rudolph and Agapie [24] studied multiobjective optimizers with respect to their limit behavior. They aimed to find whether a particular algorithm converges if the number of iterations goes to infinity. In the single objective case, the working of EAs have been analyzed for many analytical functions - linear functions [15]; unimodal functions [13]; and quadratic functions [18]. Some recent work has been done on sorting and shortest-path problems by recasting them as combinatorial problems [32]. A study is done to evaluate the blackbox complexity of problems too [33]. All the above work used a base-line simple (1 + 1) EA. The analysis of (1 + 1) EA was done using the method of partitioning of the decision space in accordance with the complexity of the problem into fitness layers as one of the methods [15]. For all such work, the only genetic operator used was mutation. However, Jansen and Wegener [19] analyzed the effectiveness of crossover operator, and showed that crossover does help for a class of problems. For multiobjective optimization, the analysis of the asymptotic expected optimization time has been started by Laumanns et al. [10]. They presented the analysis of population-based EAs (SEMO and FEMO) on a bi-objective problem (LOTZ) with conflicting objectives. They extended this work by introducing another pseudo-boolean problem (Count Ones Count Zeros: COCZ), another algorithm Greedy Evolutionary Multiobjective Optimizer (GEMO), and scaling the LOTZ and COCZ problems to larger number of decision variables and of objectives [11]. A similar analysis was performed by Giel [25] and Thierens [26] on another bi-objective problem (Multiobjective Count Ones: MOCO) and the quadratic function that we use for our algorithm. These authors designed simple toy functions to understand the behavior of simple EAs for multiobjective problems. In [12] the authors solve a special instance of the multiobjective knapsack problem. 2.2 Algorithms Analyzed The single objective optimizer basically yields a single optimal solution so (1 + 1) EAs have been successfully used and analyzed for different functions. However, in the multiobjective case, an optimizer should return a set of incomparable or equivalent

4 Running Time Analysis of a Multiobjective Evolutionary Algorithm 115 solutions. Hence a population-based EA is preferred. For this purpose, Laumanns et al. proposed a base-line population based EA called Simple Evolutionary Multiobjective Optimizer (SEMO). Another strategy used is a multi-start variant of (1 + 1) EA [11]. These algorithms have an unbounded population. Individuals are added or removed from the population based on some selection criterion. Laumanns et al. introduced two other variants of SEMO called FEMO (Fair EMO) and GEMO (Greedy EMO) which differ in their selection schemes. The algorithms do not have any defined stopping criterion and are run till the desired representative approximate set of the Pareto-front is in the population [11]. There is another group of algorithms which use an explicit or implicit archiving strategy to store the best individuals obtained so far. This approach has proven to be very effective in finding the optimal Pareto-front at much reduced computational cost, e.g., NSGA-II [34], PAES [35], PCGA [36] and SPEA2 [37]. Also, many real-world problems have effectively been solved using such a strategy. However, there exists no analysis of such algorithms. In this work, we propose and use an archive-based EA. Another issue in archive-based EAs is the size of the archive and the mating pool. If we restrict the number of individuals used for mating in the population to a constant, the expected waiting time till the desired individual is selected for mutation, is considerably reduced. But, for such an algorithm an efficient selection strategy to choose the proper individuals from the archive to the population needs to be devised. This is further discussed while formulating and analyzing the REMO algorithm 3 Basic Definitions In the multiobjective optimization scenario there are m incommensurable and often conflicting objectives that need to be optimized simultaneously. We formally define some terms below which are important from the perspective of MOEAs. We follow [3, 23 25, 36, 38] for some of the definitions. Definition 1. Multiobjective Optimization Problem (MOP): A general Multiobjective Optimization problem includes a set of n decision variables x =(x 1,x 2,..., x n ), asetofm objective functions F = {f 1,f 2,..., f m } and a set of k constraints C = {c 1,c 2,..., c k }. The objectives and the constraints are functions of the decision variables. The goal is to: Maximize/Minimize: F (x) ={f 1 (x),f 2 (x),..., f m (x)} subject to satisfaction of the constraints: C(x) ={c 1 (x),c 2 (x),..., c k (x)} 0 for X =(x 1,x 2,..., x n ) Y F =(f 1,f 2,..., f m ) G where X is the decision vector, F is the objective vector, Y denotes the decision space and G is a function space. The m objectives may be mutually conflicting in nature. In some formalizations the k constraints defined above are treated as objective functions, thus, making the problem constraint-free, and vice-versa the objectives may be treated as constraints to reduce the dimensionality of the objective-space.

5 116 Rajeev Kumar and Nilanjan Banerjee Definition 2. Quasi Order, Partial Order: A binary relation on a set F is called a quasi order if it is both reflexive and transitive. The pair (F, ) is called a partially ordered set if it is an antisymmetric quasi order. The Pareto dominance relationship in multiobjective evolutionary algorithms are partial orders (posets). The reason for the Pareto dominance relation to be a partial order is that there might be a number of individuals in the population which are mutually incomparable or equivalent to each other. An ordering is not defined for them. Definition 3. q : Let Y be the decision space and let (G, ) be a poset of objective values. Let f : Y G be a mapping. Then f induces a quasi-order q on Y (a set of binary vectors) by the following definitions for a minimization problem: x q y iff f(x) f(y) x = q y iff f(x) =f(y) x q y iff x q y x = q y. The above definition introduces the concept of Pareto-dominance and Pareto optimality. Pareto-dominance can be defined as follows: In a maximization problem of m objectives o 1,o 2,..., o m, an individual objective vector F i is partially less than another individual objective vector F j (symbolically represented by F i F j )iff( oi )(f oi i f oi j ) ( o j )(f oj i <f oj oi j ),wherefi and f oj i are components of F i and F j respectively. Then F j is said to dominate F i. If an individual is not dominated by any other individual, it is said to be non-dominated. We use the notion of Pareto-optimality if F =(f 1,..., f m ) is a vector-valued objective function. Pareto dominance is formally defined in the next definition. Definition 4. Pareto Dominance: A vector f m = {f1 m,..., fk m } is said to dominate a vector f n = {f1 n,..., fj n} (denoted by f m f n )ifff n is partially less than f m, i.e., i {1,..., k},fm i f n i i {1,..., k} : f i m <fi n. Definition 5. Pareto Optimal Set: AsetA Y (where Y denotes the entire decision space) is called a Pareto optimal set iff a A: there does not exist x Y : a x. In most practical cases it is not possible to generate the entire Pareto-optimal set. This might be the case when the size of the set is exponentially large in the input size. Thus, we confine our goal to attain an approximate set. This approximate set is usually polynomial in size. Since in most cases the objective functions are not bijective, there are a number of individuals in the decision space which are mapped to the same objective function. Hence, one might define an approximate set by selecting only one individual corresponding to an objective function. This is usually done in the case of single objective optimization problem. Definition 6. Approximate Set: AsetA p A (Pareto-optimal Set) is called an approximate set if there is no individual in A p which is weakly dominated by any other member of A p.

6 Running Time Analysis of a Multiobjective Evolutionary Algorithm 117 Another strategy that might be used to attain an approximate set is to try and obtain an inferior Pareto front. Such a front may be inferior with respect to the distance from the actual front in the decision space or the objective space. If the front differs from the actual optimal front by a distance of ɛ in the objective space, then, the dominance relation is called a (1 + ɛ)-dominance. Definition 7. (1 + ɛ)-domination: For decision vectors X 1,X 2 X where X 1 = (x 11,x 12,..., x 1i ) and X 2 =(x 21,x 22,..., x 2i ), we say that X 2 (1 + ɛ)-dominates X 1, denoted by X 1 1+ɛ X 2,iff(x 1i ) (1+ɛ) f(x 2i ) for all objectives to be maximized, and f(x 2i ) (1 + ɛ) f(x 1i ) for all objectives to be minimized. The (1 + ɛ)-dominance is transitive. The optimal set created by applying the above dominance relation is called a (1 + ɛ)-approximate set. Definition 8. (1+ɛ)-Approximate Set: AsetA 1+ɛ is called a (1+ɛ)-approximate set of the Pareto set if for all elements a p in the Pareto-set there exists an element a A 1+ɛ such that a p 1+ɛ a. Definition 9. δ-sampling of the Pareto-Set: If P denotes the Pareto-optimal set, then a δ-sampling of P is a set P P, such that no two individuals in P is within a distance of δ units in the objective space (assuming some metric in the objective space). One might also attain an approximate set by taking a proper subset of the Paretooptimal set. A strategy used to get such a subset is called sampling. A δ-sampling is a special form of sampling in which no two individuals are within a distance of δ in the objective space. The reason is to reduce the output length of the algorithm and reduce the problem to P -space. Definition 10. Running Time of an EA: The running time of an EA searching for an approximate set is defined as the number of iterations of the EA loop until the population is an approximate set for the considered problem. 4 Algorithm The algorithm suggested in this paper uses a restricted mating pool or population P of only two individuals and a separate archive A for storing all other points that are likely to be produced during a run of the algorithm. The two individuals to be chosen are selected based on a special function called Handler. With a probability of 1 2 a function Fit(x, P A) =H(x) is evaluated where H(x) is the number of hamming neighbors of x in P A. The two individuals with the smallest Fit-values are selected into the population P and the rest are transferred to the archive A. Such a selection strategy assures that we select an individual for mating that has a higher probability of producing a new offspring in the next run of the algorithm. Such a selection strategy has been found to improve the expected running time for simple functions like LOTZ. The intuition behind such a selection scheme is that for simple functions whose optimal set is uniformly distributed over the front and whose individuals are Hamming neighbors of each other. However, for the other half of the cases the handler function selects the two individuals at random. This is similar to the selection mechanism in SEMO [10]. This

7 118 Rajeev Kumar and Nilanjan Banerjee is done because in certain functions it is probable that if the two individuals selected are always those with the largest hamming distance, the algorithm might not be able to explore all the bit vectors in the optimal set. This happens for harder problems like the 0-1 knapsack. The algorithm takes ɛ (ɛ 0) as an input parameter and produces as its output a (1 + ɛ)-approximate set of the Pareto-optimal set. The algorithms in its main loop creates an individual uniformly at random and adds the individual into the population if it is not weakly dominated by any other individual in the population and it is not dominated by any other individual. All those individuals which are dominated by the new individual and if the new individual does not dominate the individual in the P and A are removed from P and A. Such a strategy is adopted so that if any individual from the (1 + ɛ)-set is created it is never removed from the population. Note that if ɛ =0and we aim to find the entire Pareto set then we only need to check whether there is some individual which is dominated by the newly created individual and remove it from P and A. Restricted Evolutionary Multiobjective Optimizer (REMO) 1. Input Parameter: ɛ, ɛ 0, ifwedesirea(1 + ɛ)-approximate set of the Paretooptimal set. 2. Initialize two sets P = φ and A = φ, wherep is the mating pool and A is an archive. 3. Choose an individual x uniformly at random from Y = {0, 1} n. 4. P = {x}. 5. loop 6. Select an individual y from P at random. 7. Apply mutation operator on y by flipping a single randomly chosen bit and create y. 8. P = P \{l P l y l does not (1 + ɛ)-dominate y }. 9. A = A \{l A l y l does not (1 + ɛ)-dominate y }. 10. if there does not exist z P A such that y z or f(z) =f(y ) then P = P {y }. 11. end if. 12. if cardinality of P is greater than 2 then 13. Handler Function 14. end if. 15. end loop. Handler Function 1. Generate a random number R in the interval (0, 1). 2. if R> 1 2 then Step 3 else Step For all the members of P A calculate a fitness function Fit(x, P A) =H(x) where H(x) denotes the number of hamming neighbors of x in P A. 4. Select two individuals with the minimum Fit(x, P A) values into P and put the rest of the individuals in the archive A. In case of equal Fit(x, P A) values the selection is made at random. 5. Select two individuals at random from P A.

8 Running Time Analysis of a Multiobjective Evolutionary Algorithm Analysis of REMO on Simple Functions 5.1 Leading Ones Trailing Zeros The Leading Ones (LO), Trailing Zeros (TZ) and the LOTZ problems can be defined as follows where the aim is to maximize both the objectives: LO : {0, 1} n N LO(x) = n i i=1 j=1 x j TZ : {0, 1} n N TZ(x) = n n i=1 j=i (1 x j) LOTZ : {0, 1} n N 2 LOTZ(x) = (LO(x),TZ(x)) Proposition 1. The Pareto front (optimal set of points in the objective space) for LOTZ can be represented as a set S = {(i, n i) 0 i n} and the Pareto set consists of all bit vectors belonging to the set P = { 1 i 0 n i 0 i n }[10]. Proof. The proof is the same as given in [10]. Analysis: The analysis of the function above is divided into two distinct phases. Phase 1 ends with the first Pareto-optimal point in the population P, and Phase 2 ends with the entire Pareto-optimal set in P A. We assume ɛ =0, thus, we aim to find the entire Pareto-optimal set. Theorem 2. The expected running time of REMO on LOTZ is O(n 2 ) with a probability of 1 e Ω(n). Proof. We partition the decision space into fitness layers defined as (i, j), (0 i, j n) where i refers to the number of Leading-ones and j is the number of Trailing-zeros in a chromosome. The individuals in one particular fitness layer are incomparable to each other. A parent is considered to climb up a fitness layer if it produces a child which dominates it. In Phase 1 a mutation event is considered a success if we climb up a fitness layer. If the probability of a success S, denoted by P (S) p i then the expected waiting time for a success E(S) 1 p i. For LOTZ, in Phase 1 the population cannot contain more than one individual for REMO because a single bit flip will create a child that is either dominating or is dominated by its parent and the algorithm does not accept weakly dominated individuals. The decision space is partitioned into fitness layers as defined above. Phase 1 begins with a initial random bit vector in P. Let us assume that after T iterations in Phase 1 the individual A(i, j) is in the population P. The individual can climb up a fitness layer (i, j) by a single bit mutation if it produces the child (i +1,j) or (i, j +1). The probability of flipping any particular bit in the parent is 1 n, thus the probability associated with such a transition is 2 n. The factor of 2 is multiplied because we could either flip the leftmost 0 or the rightmost 1 for a success. Therefore, the expected waiting time for such a successful bit flip is at most n 2. If we assume that Phase 1 begins with the worst individual (0, 0) in the population then algorithm would require at most n successful mutation steps till the first Pareto-optimal point is found. Thus, it takes i=n i=1 n 2 = n2 2 expected number of steps for the completion of Phase 1.

9 120 Rajeev Kumar and Nilanjan Banerjee To prove that the above bound holds with an overwhelming probability let us consider that the algorithm is run for n 2 steps. The expected number of successes for these n 2 steps is at least 2n.IfS denotes the number of successes, then by Chernoff s bounds: P [S (1 1 2 ) 2n] =P [S n] e n 4 = e Ω(n) Hence, the above bound for Phase 1 holds with a probability of 1 e Ω(n) which is exponentially close to 1. Phase 2 begins with an individual of the form I =(i, n i) in P. A success in Phase 2 is defined as the production of another Pareto-optimal individual. The first successful mutation in Phase 2 leads to production of the individual I +1 =(i +1,n i 1) or I 1 =(i 1,n i +1))in the population P. The probability of such a step is given by 2 n. Thus, the waiting time till the first success occurs is given by n 2. If we assume that after the first success I and I 1 are in P (without loss of generality), then the Paretooptimal front can be described as two paths from 1 i 1 0 n i+1 to 0 n and 1 i 0 n i to 1 n. At any instance of time T, let the individuals in P be represented by L =(l, n l) and K =(k, n k) where 0 k<l n. As the algorithm would have followed the path from (i 1,n i +1)to (k, n k) and (i, n i) to (l, n l) to reach to the points L and K, it is clear that at time T all the individuals of the form S =(j, n j) with l<j<khave already been found and form a part of the archive A. Moreover,the handler function, assures that L and K are farthest apart as far as Hamming distance is concerned. At time T the probability of choosing any one individual for mutation is 1 2. Let us assume, without loss of generality, that the individual selected is (k, n k). The flipping of the left most 0 produces the individual K +1 =(k +1,n k 1) and the flipping of the rightmost 1 produces the individual K 1 =(k 1,n k +1). Since, the algorithm does not accept weakly dominated individuals and K +1 is already in A, the production of K 1 can only be considered as a success. Thus the probability of producing another Pareto-optimal individual at time T is 1 2n. Thus, the expected waiting time of producing another Pareto-optimal individual is at most 2n. Since, no solutions on the Pareto-optimal front is revisited in Phase 2, it takes a maximum of n +1steps for its completion. The special case hold when the individual 0 n or 1 n appears in the population. Such individuals represent the end points of the Pareto-front and their mutation do not produce any individual that can be accepted. Moreover, such individuals always form a part of the population P since they have to be a part of the pair of individuals which have the maximum Hamming distance. However, if such an individual is a part of P the probability bound still holds as the probability of choosing the right individual (the individual which is not 0 n or 1 n ) for mutation is still 1 2 and the probability of successful mutation is 1 2n and expected running time bound holds. Therefore, REMO takes O(n 2 ) for Phase 2. Now, we consider Phase 2 with 4n 2 steps. By arguments similar to Phase 1, it can be shown by Chernoff s bounds that the probability of the number of successes in Phase 2 being less than n,ise Ω(n). Altogether considering both the Phases, REMO takes n 2 steps to find the entire Pareto-optimal set for LOTZ. For the bound on the expected time we have not assumed anything about the initial population. Thus, we notice that the above bound on the probability holds for the next n 2 steps. Since the lower bound on the probability that the algorithm will find the entire

10 Running Time Analysis of a Multiobjective Evolutionary Algorithm 121 Pareto set is more than 1 2 (in fact exponentially close to 1) the expected number of times the algorithm has to run is bounded by 2. Combining the results of both the phases 1 and 2 yields the bounds in the theorem. Comments: The above bound only considers the number of iterations that the algorithm needs to find the Pareto set. However, if all the evaluations of the loop is considered, the handler function can take a time O(n 2 ) to find the pair of individuals with the least Fitvalue and hence, the entire algorithm may take time of O(n 4 ). 5.2 Quadratic Function We use a continuous quadratic function which has been widely used in empirical analysis of multiobjective evolutionary optimizers and adapt it to the discrete boolean decision space exactly as done in [25]. The function in the continuous decision space is ((x a) 2, (x b) 2 ) where the aim is the minimization of both the objectives. The function in the discrete boolean decision space is described in the following manner exactly as in [25]: QF : {0, 1} n N 2. if x = n i=1 x i QF :( x a) 2, ( x b) 2 ) The idea is to just test the efficiency of REMO in solving problems like those of QF and in fact the analysis follows a line similar to that of [25]. Proposition 2. Without loss of generality we assume that a>b. The Pareto-optimal front of QF can be represented by the set F = {(i 2, (i (b a)) 2 ) a i b}. The Pareto-optimal points of QF are those bit vectors where a x b. Proof. The proof is the same as given in [25]. Theorem 2. The running time of REMO on QF is O(n log n) for any value of a and b. (We assume ɛ =0, thus we aim to find the entire Pareto-optimal set.) Proof. We partition the analysis into two phases. The analysis turns out to be very similar to that done in [25] but it is much simpler due to the local mutation operator used in REMO. Phase 1 ends with the first Pareto-optimal point in P and the second phase continues till all the Pareto-optimal bit vectors are in P A. It can be proven that in Phase 1 there can be a maximum of 1 (similar to [25] individuals in P A. Thus, the archive A is empty. This is because a single bit mutation of a parent with x <aor x >bwillproduce an individual which is dominated by or dominates its parent. We partition the decision space into sets with individuals having the same number of ones. Let us consider a bit vector represented as I d where d represents the number of ones in the individual. A single bit mutation of I d is considered to be a success if the number of ones increases (decreases) when d<a(d >b).

11 122 Rajeev Kumar and Nilanjan Banerjee Therefore, a success S requires the flipping of any one of the d 1-bits (n d 0-bits) when d<a(d >b). The probability of a successful mutation P (S) = d n d n (or n ).The expected waiting time of S, givenbye(s) n d (or n n d ). Hence, the total expected time till the first Pareto optimal point arrives in the population is at most n d=1 n d = nh n = n log n + Θ(2n) =O(n log n) (where H n is the n th Harmonic number) by the linearity of expectations. Phase 2 starts with the assumption that b a>1or else there would be no second phase. The number of individuals in the population is bounded by 2. The selection mechanism ensures that they are the bit vectors that are most capable of producing new individuals. The Pareto-front can be visualized as a path of individuals with number of ones varying from a to b or b to a. Let us represent any individual with a< x <b as I k where k represents the number of ones in the bit vector. Such a bit vector can be produced either by an individual with k +1ones or k 1 ones. The associated probability for such a successful mutation is at least k+1 n k+1 2n and 2n respectively. Hence, the expected waiting time till the Ik th Pareto optimal point is in the population (assuming that its parent is in the population) is E(I k ) 2n k+1 and 2n n k+1 for the two cases above. Thus, the total expected time till all the Pareto points are in P A is at most b k=a E(I k) b k=a 2n k+1 b a k=0 2n k+1 = 2nH b a where H n stands for the n th harmonic number. Therefore, the expected time for Phase 2 is at most 2ne log(b a) +θ(2ne) = O(n log(b a)). Altogether both Phases take a total of O(n log n + n log(b a)) time. Since a and b can have a maximum value of n the running time for REMO on QF is O(n log n) which is the same as proven in [25]. 6 Linear Functions and Knapsack Problem 6.1 Linear Functions A bi-objective linear function is: F (x) = (f 1 (x) = n i=1 w ix i,f 2 (x) = n i=1 w i x i) where w i > 0,w i > 0 The aim of a bi-objective problem may be to maximize or minimize both the objectives; in such a case the problem reduces to a single objective one. In our study, therefore, we take the case of simultaneously maximizing f 1 and minimizing f 2. Thus, the problem is formulated with two mutually conflicting objectives. In this section, we show that for the function F (x) the number of Pareto-optimal points can range from n+1 to 2 n. Thus, for any arbitrary values of the weights w and w the Pareto-optimal set can be exponential in n. Throughout this section we investigate the case where the bits of the individuals are arranged in their decreasing value of w w. Thus, w1 w w2 1 w... wn 2 w. n Lemma 1. For any bi-objective linear function F (x) =(f 1,f 2 )theseta 1 = {1 i 0 n i } where 0 i n represents a set of Pareto-optimal points.

12 Running Time Analysis of a Multiobjective Evolutionary Algorithm 123 Proof. Let us consider an individual K in the decision space (which does not belong to A 1 ) and an individual K A 1 which is of the form (1 k 0 n k )for0 k n. Ifthe set of 1-bits of K is a subset of the set of 1-bits in K it is clear that K and K are incomparable. However, if the set of 1-bits of K is not a subset of K, lets denote the set of common bit positions that are set to one in both K and K with x 1 = i S w i.lets 1 denote the set of bit positions that are set to one in K but not in K and y 1 = i S 1 w i. If the individual K has to dominate K, in the best case f 2 (K ) is at most f 2 (K).Since all the bits are arranged in the decreasing order of wi w, f i 1 (K ) is at most x 1 + w k+1 w y 1 k+1 y 1.Now,f 1 (K) is at least x 1 + w k w y 1. Hence, f 1 (K) >f 1 (K ). Therefore, k x 1 + w k w k K cannot dominate K. Now we need to prove that two individuals in A 1 are mutually incomparable to each other. Let us consider another individual I =1 i 0 n i in A 1 where 0 i n. If i<k, f 1 (I) <f 1 (K) and f 2 (I) <f 2 (K), implying that I and K are incomparable (by definition 2). A similar argument holds for i>k, hence proving the lemma. Example 1. Let us consider a linear function with three weights. The w of the components are W = {20, 15, 19} and the weights W are W = {8, 7, 12}. Clearly W 1 W > W2 1 W > W3 2 W. Let us consider the bit vector 110. This individual has the first 3 and second weights set to 1. Individuals whose bits set to 1 are a subset of the above individuals, for example 100, will have both w and w less than 110, and hence is incomparable to it. Individuals which are not a subset of 110, like for example 101 will be dominated by 110 or is incomparable to it. As an example, w(110) = 35, w (110) =15 and w(101) = 39, w (101) = 20, these two individuals are incomparable. This hold for any individual which is not a subset of 110. Proposition 3. The size of the Pareto-optimal set for the most general case of a linear function F (x) lies between n +1and 2 n. Proof. It is clear from lemma 1 that the lower bound on the number of Pareto-optimal individuals for F (x) is n +1. Moreover, the upper bound holds for cases where all the bit vectors are Pareto-optimal. We next show that there are examples which fit into the above bounds. Case 1: Let us consider a linear function such that w 1 >w 2 >w 3 >... > w n and w 1 <w 2 <w 3 <... < w n. Each Pareto-optimal point is of the form X =1 i 0 n i where 0 i n. It is clear that individuals of the form X represent a Pareto-optimal solution because it contains the i largest weights of f 1 and the i smallest weights of f 2. Flipping the left-most 0-bit of X to 1 or the right-most 1-bit to 0 creates an individual which is incomparable to X. Moreover, any individual with a 0 followed by a 1 cannot be Pareto-optimal as it can be improved in both objectives by simply swapping the bits. The Pareto-optimal set of such a function thus contains n +1individuals. Case 2: For the other extreme case, let us consider a linear function for which w1 w = 1 =... = wn w and w n 1 >w 2 >w 3 >... > w n. It is clear that for such a w 2 w 2 = w3 w 3

13 124 Rajeev Kumar and Nilanjan Banerjee function all the points in the decision space {0, 1} n are Pareto-optimal. Thus, the total number of Pareto points for this case is 2 n. Consequently, for any randomized algorithm the expected runtime to find the entire Pareto-optimal set for the above case of bi-objective linear functions is Ω(2 n ). 6.2 Knapsack Problem Next, we show that the above problem of the conflicting objectives for a linear function can be interpreted as the 0-1 Knapsack problem. Definition Knapsack Problem: The knapsack problem with n items is described by the knapsack of size b and three sets of variables related to the items: decision variables x 1,x 2,..., x n ; positive weights W 1,W 2,..., W n ; and profits P 1,P 2,..., P n ; where, for each 1 i n, x i is either 0 or 1. The W i and P i represent the weight and profit, as integers, of the i th item respectively. The single-objective knapsack problem can be formally stated as: Maximize P = n subject to n i=1 P ix i i=1 W ix i b, where x i =0or 1 In order to recast the above single-objective problem along with one constraint on weights of the items into a bi-objective problem, we use the formulation similar to the linear function described above. Thus, a bi-objective knapsack problem of n items with two conflicting objectives (maximizing profits and minimizing weights) is formulated as: Maximize P = n j=1 P jx j and Minimize W = n j=1 W jx j Therefore, if the set of items is denoted by I, the aim is to find all sets I i I, such that for each I j there is no other set which has a larger profit than profit (I j )for the given weight bound W (I j ). Thus, it is equivalent to finding the collections of items with the maximum profit in a knapsack with capacity W (I j ). The 0 1 knapsack problem, in the optimization form above is NP-complete. In the following section we formalize a (1 + ɛ)- approximate set for the above knapsack problem. We work under the assumption that the items are arranged in a strictly decreasing value of P W. Lemma 2. Let us define a set Xj i as the set of i most efficient items (efficiency defined by P W ) among the j smallest weight items if the sum of the weights of the j items is less than the (j +1) st item. Let A 1 be the set of all such Xj i and the following constraint be imposed on weights of the items: Xj i,if{ij 1,..., Ij i } represents the set of items in Xj i,thenw I i+1 <ɛ k=i k=1 W I k for i<j.now,ifa 2 represents a singleton set, {0 i 10 n (i+1) } where the 1 is set at the position of the lightest item. Then, A 1 A 2 represents a(1 + ɛ)-approximation of the Pareto-optimal set for the knapsack problem.

14 Running Time Analysis of a Multiobjective Evolutionary Algorithm 125 Example 2. Let us consider a knapsack with three items. The profits of the items is given by P = {40, 20, 10} and the weights by W = {20, 11, 6}. The knapsack satisfies the constraint on the weights (given in Lemma 2) for ɛ =0.6. X 1 is a trivial case as it contains the lightest item. For X i 2,1 i 2, we consider the second and the third items. Clearly W 3 <ɛ W 2 and hence it satisfies the constraint. For X i 3, 1 i 3 we have all the items. Since, W 2 <ɛ W 1 and W 3 <ɛ (W 1 + W 2 ), the knapsack weights satisfy the given constraint. Proof. Let (P o,w o ) represent any arbitrary Pareto-optimal solution for the knapsack problem. We need to prove, that corresponding to such a solution we can always find a solution (P,W ) in A 1 A 2 such that (P o,w o ) 1+ɛ (P,W ).IfW o is the item with the smallest weight then the element in set A 2 weakly dominates (P o,w o ) and hence (1 + ɛ)-dominates (P o,w o ). Now, let us consider a more general case. Let π define the permutation of the items which reorders them according to their weights. Corresponding to any W o, we aim to find an item I π(j) such that W π(j) (1+ɛ) W o < W π(j+1). If such an item cannot be found then j = n +1. It is clear that for this j +1,anXj+1 i A 1, a set of items I j+1 1,I j+1 2,..., I j+1 i (i can be equal to j), can always be found such that k=i k=1 W I k (1 + ɛ) W o < k=i+1 k=1 W Ik. We claim that W o k=i k=1 W I k (1 + ɛ) W o. This holds because of the constraint on the weights imposed in the lemma. It is clear from the constraint that if W Ii+1 ɛ i k=1 W I k and adding of the (i +1) st item into the knapsack increases the weight of the knapsack above (1+ɛ) W o, the sum of the weights of the items in Xj+1 i is at least W o. Since, the sum of the weights in the set Xj+1 i obeys the weight bound (1 + ɛ) W o, with respect to the weights, (P o,w o ) 1+ɛ (P,W ). Now, we know that W W o.lets c denote the items common in both the solution (P,W ) and (P o,w o ) and S un denote the set of items in the solution (P,W ) and not in (P o,w o ).LetA = I m S c P Im and B = I m S uc W Im.Now,W o <W π(j+1), therefore (P o,w o ) contains items in the set X j+1 (which is the set of the first j +1lightest items). Since Xj+1 i is the set of the i most efficient items in X j+1 (efficiency defined as P W ), P o A + B Pi+1 W i+1 A + B Pi W i P, thus proving that (P o,w o ) 1+ɛ (P,W ), hence the lemma. In the next lemma, we extend the results obtained in lemma 2 to formalize a (1+ɛ)- approximate set for any knapsack. Lemma 3. Let I = {I 1,I 2,..., I n } represent the set of items of the knapsack. We partition the set I into m blocks (1 m n), B 1,B 2,..., B m which satisfy the following conditions: 1. Each block consists of a set of items and all pairs of blocks are mutually disjoint. However, B 1 B 2... B m = I. 2. The items in each block satisfy the weight constraint described in lemma 3 and are arranged in their strictly decreasing Pi W i ratio in the block. Sets A 1 and A 2, similar to those defined in lemma 2 are defined for every block B m. Let A m = A m 1 Am 2 for the mth block. If S denotes the set of items formed by taking

15 126 Rajeev Kumar and Nilanjan Banerjee one set from every A m, then, the collection of all such S sets, represented by S comp, denotes a (1 + ɛ)-approximation of the Pareto-optimal set for any knapsack problem. If m = n the set reduces to a power set of the item-set. Example 3. Let us consider a knapsack of four items, {I 1,I 2,I 3,I 4 }. The profits of the items are given by P = {40, 20, 10, 15} and the weights are given by W = {20, 11, 6, 40}. The items of the above knapsack can be divided into blocks B 1 = {I 1,I 2,I 3 } and B 2 = {I 4 }. The block B 1 satisfies the constraint as it is the same set of items described in example 2 and block B 2 satisfies the constraint trivially as it has only one item. Hence, the blocks describe a valid partitioning. Proof. Let us consider any Pareto-optimal collection of items U with objective values (P o,w o ). We aim to prove that corresponding to every U we can find a solution V of the form S (defined above) which (1 + ɛ)-dominates U. We partition both solutions U and V into blocks of items as defined above. Let us consider a particular block of items B i, and denote the set of items in U and V in the block as Bi U and Bi V respectively. Since, the block consists of items which satisfy all the conditions of lemma 3, the solution represented by Bi U (1 + ɛ)-dominates Bi V (irrespective of whether BU i represents a Pareto-optimal solution for the items in B i ). It is clear that the above argument holds for every block. Now, weight(u) = m i=1 W (B i) and profit(u)= m i=1 P (B i). Since, for every block B i, weight(b i (V )) < (1+ɛ) weight(b i (U)) and(1+ɛ) profit(b i (V )) > profit(b i (U)), weight(v ) < (1 +ɛ) weight(u) and(1+ɛ) profit(v ) > profit(u) by a simple summation of the profits and weights of items in every block. Therefore, U 1+ɛ V, proving the lemma. Lemma 4. The total number of individuals in the (1 + ɛ)-approximate set (defined in Lemma 3) is upper bounded by ( n m )2m,wherem is the number of blocks (m defined in lemma 3) and n is the number of items in the knapsack, if n m. Proof. Let the number of individuals in the k th block be n k. The number of sets of the form Xj i in the kth block is of the order O(n 2 k ). Thus, the total number of sets possible for all the blocks is upper bounded by (n 1 n 2... n m ) 2 if m is not a constant. Now, (n 1 n 2... n m ) 1 m (Arithmetic mean Geometric Mean). Thus, m i=1 ni m i=1 ni m 2m ) (n 1 n 2... n m ) 2 ( m = O(( n m )2m ),ifm n. However,ifn = m, by the partitioning described in Lemma 3, all the bit vectors represent the (1 + ɛ)- approximate set for the knapsack. Hence, in such a case the total number of individuals in the set is 2 n. 7 Analysis of REMO on Knapsack Problem Lemma 5. If P knap is the sum of all the profits of the items in the knapsack, then the total number of individuals in the population and archive is at most P knap +1.

16 Running Time Analysis of a Multiobjective Evolutionary Algorithm 127 Proof. It is clear that at any time the population and archive consists of individuals which are incomparable to each other as in the case of SEMO. We aim to prove that any two individuals of the population and archive will have distinct profit values. We try to prove the claim by contradiction. Let us assume that there are two individuals a and b in P A which have the same profit values. As the algorithm does not accept any weakly dominated individuals either weight(a) > weight (b) or weight(a) < weight(b). However, this contradicts our initial assumption that the population consists of individuals which are incomparable to each other. Hence, a and b have distinct profit values. As all the items have integer profits and the total profit can be zero, the total number of individuals are bounded by the sum of the profits P knap +1. Theorem 1. The expected running time for REMO to find a (1 + ɛ)-approximate set of the Pareto-front (formalization given in lemma 3) for any instance of the knapsack problem with n items is O( n2m+1 P knap m ) where m is the number of blocks into which 2m+1 the items can be divided (blocks are defined in lemma 3) and P knap refers to the sum of the profits of the items in the knapsack and n m. Moreover, the above bound holds with an overwhelming probability. It is worth noting that the expected waiting time is a polynomial in n if the sum of the profits and the number of partitions m is a polynomial in n, which in turn depends on the ɛ value. It is difficult to derive any general relationship between ɛ and m, since it depends on the distribution of the weights. However, if the ɛ is large then the number of items in a particular block is likely to be large and hence the number of blocks will be small and the running time can be polynomial in n. Proof. We divide the analysis into two phases. The first phase continues till 0 n is in the population or the archive. Phase 2 continues till all the vectors in set S is in P A.(We take ɛ as an input.) In the first phase, the aim is to have a all 0s string in the population. We partition the decision space into fitness layers. A fitness layer i is defined as a solution which has the i smallest weight items in the knapsack. At any point of time in phase 1, there is a maximum of one solution Z which has the i smallest weights in the knapsack. Removal of any one of these items from Z reduces the weight of the knapsack and hence produces a solution which is accepted. With a probability of 1 2 (as in the handler function), an individual is selected from the archive at random. Therefore, 1 the probability of selection of Z for mutation is 4(P knap +1) (P knap +1) is the bound on the population size, by lemma 5). If we flip the 1-bit corresponding to the heaviest item in Z (which occurs with a probability of 1 n ), the mutated individual is accepted. 1 Thus, the probability of producing an accepted individual is 4n(P knap +1). Therefore, the expected waiting time till a successful mutation occurs is at most 4n(P knap +1). Since, a maximum of n successes in Phase 1 assures that 0 n is in the population, the expected time for completion of Phase 1 is 4n 2 (P knap +1). Therefore, Phase 1 ends in O(n 2 P knap ) time. If we consider 8n 2 (P knap +1)steps the expected number of successes is at least 2n. By Chernoff s bound the probability of number of successes being less than n is at most e Ω(n). The second phase starts with the individual 0 n in the population. An individual that is required to be produced can be described as a collection of items I coll = C 1 C 2

17 128 Rajeev Kumar and Nilanjan Banerjee... C m,where,c k is either Xj i or one item (of the smallest weight) in the kth block. If Xj i k refers to the set X i j in the k th block, it is clear to see that H(0 n,xj 1 k )=1,where H refers to the Hamming distance. Thus, by a single bit flip of 0 n we can produce a new desired point. Since, the algorithm accepts bit vectors which are (1+ɛ)-approximations of the Pareto-optimal points, the point generated will be taken into the population and will never be removed. It is also clear that H(X i 1 k j,x ik j )=1. Hence, corresponding to every bit vector R which belongs to the (1 + ɛ)-approximate set of the Pareto front there is a bit vector in the population or the archive which by a single bit flip can produce R. Thus, from individuals like I coll, we can produce another desired individual (which belongs to the (1+ɛ)-approximate set) by a single bit flip. With a probability of 1 2 (Handler function), two individuals in the population are chosen at random. Therefore, the probability of the individual like I coll being chosen into the population is 1 2(P knap +1) (P knap +1is the bound on the population size, proven in lemma 5). The 1 probability of I coll being chosen for mutation is thus 4(P knap +1). Flipping a desired 1- bit or 0-bit in any of the m blocks of I coll will produce another desired individual. Thus, m the probability that I coll will produce another desired individual is 4n(P knap +1).Theexpected number of waiting steps for a successful mutation is thus, at most 4n(P knap+1) m. If R is the new individual produced, since, no items have equal Pi W i, there cannot be any individual in the population or the archive which weakly dominates R. Hence, R will always be accepted. As every individual in the (1 + ɛ)-approximate set can be produced from some individual in the population or the archive by a single bit mutation, the total time taken to produce the entire (1 + ɛ)-approximate set is upper bounded by 4n 2m+1 (P knap +1) m 2m+1 (total size of the (1 + ɛ)-approximate set is bounded by O(( n m )2m ) by lemma 5). If we consider a phase of 8(P knap+1)n 2m+1 m 2m+1 probability of there being less than n2m m 2m both the phases take a total of O( P knapn 2m+1 set. m 2m+1 steps, then by Chernoff s bounds, the successes is bounded by e Ω(n). Altogether ) for finding the entire (1+ɛ)- approximate For the bound on the expected time we have not assumed anything about the initial population. Thus, we notice that the above bound on the probability holds for the next P knap n 2m+1 m 2m+1 steps. Since the lower bound on the probability that the algorithm will find the entire (1 + ɛ)-approximate set is more than 1 2 (in fact exponentially close to 1) the expected number of runs is bounded by 2. Combining the results of both the phases yields the bound in the theorem. It is worth noting that for cases when the value of the number of partitions into which the set of items have to be partitioned is a constant the expected running time is a polynomialin n. Comments: An important point to note in the analysis of the algorithm is that for the knapsack problem the selection of the individual in the Handler function is done at random. However, this occurs with a probability of 0.5. It is likely that the specialized Fit function is able to find individual which is a subset of the approximate set in time which is faster than the random selection. Note that 1 i 0 n i is a subset of the approximate so-

Expected Running Time Analysis of a Multiobjective Evolutionary Algorithm on Pseudo-boolean Functions

Expected Running Time Analysis of a Multiobjective Evolutionary Algorithm on Pseudo-boolean Functions Expected Running Time Analysis of a Multiobjective Evolutionary Algorithm on Pseudo-boolean Functions Nilanjan Banerjee and Rajeev Kumar Department of Computer Science and Engineering Indian Institute

More information

An Analysis on Recombination in Multi-Objective Evolutionary Optimization

An Analysis on Recombination in Multi-Objective Evolutionary Optimization An Analysis on Recombination in Multi-Objective Evolutionary Optimization Chao Qian, Yang Yu, Zhi-Hua Zhou National Key Laboratory for Novel Software Technology Nanjing University, Nanjing 20023, China

More information

Runtime Analysis of Evolutionary Algorithms for the Knapsack Problem with Favorably Correlated Weights

Runtime Analysis of Evolutionary Algorithms for the Knapsack Problem with Favorably Correlated Weights Runtime Analysis of Evolutionary Algorithms for the Knapsack Problem with Favorably Correlated Weights Frank Neumann 1 and Andrew M. Sutton 2 1 Optimisation and Logistics, School of Computer Science, The

More information

Running Time Analysis of Multi-objective Evolutionary Algorithms on a Simple Discrete Optimization Problem

Running Time Analysis of Multi-objective Evolutionary Algorithms on a Simple Discrete Optimization Problem Running Time Analysis of Multi-objective Evolutionary Algorithms on a Simple Discrete Optimization Problem Marco Laumanns 1, Lothar Thiele 1, Eckart Zitzler 1,EmoWelzl 2,and Kalyanmoy Deb 3 1 ETH Zürich,

More information

Runtime Analyses for Using Fairness in Evolutionary Multi-Objective Optimization

Runtime Analyses for Using Fairness in Evolutionary Multi-Objective Optimization Runtime Analyses for Using Fairness in Evolutionary Multi-Objective Optimization Tobias Friedrich 1, Christian Horoba 2, and Frank Neumann 1 1 Max-Planck-Institut für Informatik, Saarbrücken, Germany 2

More information

Running time analysis of a multi-objective evolutionary algorithm on a simple discrete optimization problem

Running time analysis of a multi-objective evolutionary algorithm on a simple discrete optimization problem Research Collection Working Paper Running time analysis of a multi-objective evolutionary algorithm on a simple discrete optimization problem Author(s): Laumanns, Marco; Thiele, Lothar; Zitzler, Eckart;

More information

Plateaus Can Be Harder in Multi-Objective Optimization

Plateaus Can Be Harder in Multi-Objective Optimization Plateaus Can Be Harder in Multi-Objective Optimization Tobias Friedrich and Nils Hebbinghaus and Frank Neumann Max-Planck-Institut für Informatik, Campus E1 4, 66123 Saarbrücken, Germany Abstract In recent

More information

TECHNISCHE UNIVERSITÄT DORTMUND REIHE COMPUTATIONAL INTELLIGENCE COLLABORATIVE RESEARCH CENTER 531

TECHNISCHE UNIVERSITÄT DORTMUND REIHE COMPUTATIONAL INTELLIGENCE COLLABORATIVE RESEARCH CENTER 531 TECHNISCHE UNIVERSITÄT DORTMUND REIHE COMPUTATIONAL INTELLIGENCE COLLABORATIVE RESEARCH CENTER 531 Design and Management of Complex Technical Processes and Systems by means of Computational Intelligence

More information

A Comparison of GAs Penalizing Infeasible Solutions and Repairing Infeasible Solutions on the 0-1 Knapsack Problem

A Comparison of GAs Penalizing Infeasible Solutions and Repairing Infeasible Solutions on the 0-1 Knapsack Problem A Comparison of GAs Penalizing Infeasible Solutions and Repairing Infeasible Solutions on the 0-1 Knapsack Problem Jun He 1, Yuren Zhou 2, and Xin Yao 3 1 J. He is with the Department of Computer Science,

More information

REIHE COMPUTATIONAL INTELLIGENCE COLLABORATIVE RESEARCH CENTER 531

REIHE COMPUTATIONAL INTELLIGENCE COLLABORATIVE RESEARCH CENTER 531 U N I V E R S I T Y OF D O R T M U N D REIHE COMPUTATIONAL INTELLIGENCE COLLABORATIVE RESEARCH CENTER 531 Design and Management of Complex Technical Processes and Systems by means of Computational Intelligence

More information

Generalization of Dominance Relation-Based Replacement Rules for Memetic EMO Algorithms

Generalization of Dominance Relation-Based Replacement Rules for Memetic EMO Algorithms Generalization of Dominance Relation-Based Replacement Rules for Memetic EMO Algorithms Tadahiko Murata 1, Shiori Kaige 2, and Hisao Ishibuchi 2 1 Department of Informatics, Kansai University 2-1-1 Ryozenji-cho,

More information

Evolutionary Algorithms How to Cope With Plateaus of Constant Fitness and When to Reject Strings of The Same Fitness

Evolutionary Algorithms How to Cope With Plateaus of Constant Fitness and When to Reject Strings of The Same Fitness Evolutionary Algorithms How to Cope With Plateaus of Constant Fitness and When to Reject Strings of The Same Fitness Thomas Jansen and Ingo Wegener FB Informatik, LS 2, Univ. Dortmund, 44221 Dortmund,

More information

TECHNISCHE UNIVERSITÄT DORTMUND REIHE COMPUTATIONAL INTELLIGENCE COLLABORATIVE RESEARCH CENTER 531

TECHNISCHE UNIVERSITÄT DORTMUND REIHE COMPUTATIONAL INTELLIGENCE COLLABORATIVE RESEARCH CENTER 531 TECHNISCHE UNIVERSITÄT DORTMUND REIHE COMPUTATIONAL INTELLIGENCE COLLABORATIVE RESEARCH CENTER 531 Design and Management of Complex Technical Processes and Systems by means of Computational Intelligence

More information

TECHNISCHE UNIVERSITÄT DORTMUND REIHE COMPUTATIONAL INTELLIGENCE COLLABORATIVE RESEARCH CENTER 531

TECHNISCHE UNIVERSITÄT DORTMUND REIHE COMPUTATIONAL INTELLIGENCE COLLABORATIVE RESEARCH CENTER 531 TECHNISCHE UNIVERSITÄT DORTMUND REIHE COMPUTATIONAL INTELLIGENCE COLLABORATIVE RESEARCH CENTER 531 Design and Management of Complex Technical Processes and Systems by means of Computational Intelligence

More information

On the Impact of Objective Function Transformations on Evolutionary and Black-Box Algorithms

On the Impact of Objective Function Transformations on Evolutionary and Black-Box Algorithms On the Impact of Objective Function Transformations on Evolutionary and Black-Box Algorithms [Extended Abstract] Tobias Storch Department of Computer Science 2, University of Dortmund, 44221 Dortmund,

More information

Computational Aspects of Aggregation in Biological Systems

Computational Aspects of Aggregation in Biological Systems Computational Aspects of Aggregation in Biological Systems Vladik Kreinovich and Max Shpak University of Texas at El Paso, El Paso, TX 79968, USA vladik@utep.edu, mshpak@utep.edu Summary. Many biologically

More information

UNIVERSITY OF DORTMUND

UNIVERSITY OF DORTMUND UNIVERSITY OF DORTMUND REIHE COMPUTATIONAL INTELLIGENCE COLLABORATIVE RESEARCH CENTER 531 Design and Management of Complex Technical Processes and Systems by means of Computational Intelligence Methods

More information

Efficient Non-domination Level Update Method for Steady-State Evolutionary Multi-objective. optimization

Efficient Non-domination Level Update Method for Steady-State Evolutionary Multi-objective. optimization Efficient Non-domination Level Update Method for Steady-State Evolutionary Multi-objective Optimization Ke Li, Kalyanmoy Deb, Fellow, IEEE, Qingfu Zhang, Senior Member, IEEE, and Qiang Zhang COIN Report

More information

When to Use Bit-Wise Neutrality

When to Use Bit-Wise Neutrality When to Use Bit-Wise Neutrality Tobias Friedrich Department 1: Algorithms and Complexity Max-Planck-Institut für Informatik Saarbrücken, Germany Frank Neumann Department 1: Algorithms and Complexity Max-Planck-Institut

More information

REIHE COMPUTATIONAL INTELLIGENCE S O N D E R F O R S C H U N G S B E R E I C H 5 3 1

REIHE COMPUTATIONAL INTELLIGENCE S O N D E R F O R S C H U N G S B E R E I C H 5 3 1 U N I V E R S I T Ä T D O R T M U N D REIHE COMPUTATIONAL INTELLIGENCE S O N D E R F O R S C H U N G S B E R E I C H 5 3 1 Design und Management komplexer technischer Prozesse und Systeme mit Methoden

More information

Lecture 4: An FPTAS for Knapsack, and K-Center

Lecture 4: An FPTAS for Knapsack, and K-Center Comp 260: Advanced Algorithms Tufts University, Spring 2016 Prof. Lenore Cowen Scribe: Eric Bailey Lecture 4: An FPTAS for Knapsack, and K-Center 1 Introduction Definition 1.0.1. The Knapsack problem (restated)

More information

REIHE COMPUTATIONAL INTELLIGENCE S O N D E R F O R S C H U N G S B E R E I C H 5 3 1

REIHE COMPUTATIONAL INTELLIGENCE S O N D E R F O R S C H U N G S B E R E I C H 5 3 1 U N I V E R S I T Ä T D O R T M U N D REIHE COMPUTATIONAL INTELLIGENCE S O N D E R F O R S C H U N G S B E R E I C H 5 3 1 Design und Management komplexer technischer Prozesse und Systeme mit Methoden

More information

Limitations of Algorithm Power

Limitations of Algorithm Power Limitations of Algorithm Power Objectives We now move into the third and final major theme for this course. 1. Tools for analyzing algorithms. 2. Design strategies for designing algorithms. 3. Identifying

More information

Effects of the Use of Non-Geometric Binary Crossover on Evolutionary Multiobjective Optimization

Effects of the Use of Non-Geometric Binary Crossover on Evolutionary Multiobjective Optimization Effects of the Use of Non-Geometric Binary Crossover on Evolutionary Multiobjective Optimization Hisao Ishibuchi, Yusuke Nojima, Noritaka Tsukamoto, and Ken Ohara Graduate School of Engineering, Osaka

More information

NP Completeness and Approximation Algorithms

NP Completeness and Approximation Algorithms Chapter 10 NP Completeness and Approximation Algorithms Let C() be a class of problems defined by some property. We are interested in characterizing the hardest problems in the class, so that if we can

More information

DRAFT -- DRAFT -- DRAFT -- DRAFT -- DRAFT --

DRAFT -- DRAFT -- DRAFT -- DRAFT -- DRAFT -- Towards an Analytic Framework for Analysing the Computation Time of Evolutionary Algorithms Jun He and Xin Yao Abstract In spite of many applications of evolutionary algorithms in optimisation, theoretical

More information

CS 350 Algorithms and Complexity

CS 350 Algorithms and Complexity 1 CS 350 Algorithms and Complexity Fall 2015 Lecture 15: Limitations of Algorithmic Power Introduction to complexity theory Andrew P. Black Department of Computer Science Portland State University Lower

More information

Containment restrictions

Containment restrictions Containment restrictions Tibor Szabó Extremal Combinatorics, FU Berlin, WiSe 207 8 In this chapter we switch from studying constraints on the set operation intersection, to constraints on the set relation

More information

A Brief Introduction to Multiobjective Optimization Techniques

A Brief Introduction to Multiobjective Optimization Techniques Università di Catania Dipartimento di Ingegneria Informatica e delle Telecomunicazioni A Brief Introduction to Multiobjective Optimization Techniques Maurizio Palesi Maurizio Palesi [mpalesi@diit.unict.it]

More information

CS 350 Algorithms and Complexity

CS 350 Algorithms and Complexity CS 350 Algorithms and Complexity Winter 2019 Lecture 15: Limitations of Algorithmic Power Introduction to complexity theory Andrew P. Black Department of Computer Science Portland State University Lower

More information

Genetic Algorithms: Basic Principles and Applications

Genetic Algorithms: Basic Principles and Applications Genetic Algorithms: Basic Principles and Applications C. A. MURTHY MACHINE INTELLIGENCE UNIT INDIAN STATISTICAL INSTITUTE 203, B.T.ROAD KOLKATA-700108 e-mail: murthy@isical.ac.in Genetic algorithms (GAs)

More information

Quad-trees: A Data Structure for Storing Pareto-sets in Multi-objective Evolutionary Algorithms with Elitism

Quad-trees: A Data Structure for Storing Pareto-sets in Multi-objective Evolutionary Algorithms with Elitism Quad-trees: A Data Structure for Storing Pareto-sets in Multi-objective Evolutionary Algorithms with Elitism Sanaz Mostaghim 1 and Jürgen Teich 2 1 Electrical Engineering Department University of Paderborn,

More information

A Lower Bound Analysis of Population-based Evolutionary Algorithms for Pseudo-Boolean Functions

A Lower Bound Analysis of Population-based Evolutionary Algorithms for Pseudo-Boolean Functions A Lower Bound Analysis of Population-based Evolutionary Algorithms for Pseudo-Boolean Functions Chao Qian,2, Yang Yu 2, and Zhi-Hua Zhou 2 UBRI, School of Computer Science and Technology, University of

More information

CMSC 451: Lecture 7 Greedy Algorithms for Scheduling Tuesday, Sep 19, 2017

CMSC 451: Lecture 7 Greedy Algorithms for Scheduling Tuesday, Sep 19, 2017 CMSC CMSC : Lecture Greedy Algorithms for Scheduling Tuesday, Sep 9, 0 Reading: Sects.. and. of KT. (Not covered in DPV.) Interval Scheduling: We continue our discussion of greedy algorithms with a number

More information

Outline. 1 NP-Completeness Theory. 2 Limitation of Computation. 3 Examples. 4 Decision Problems. 5 Verification Algorithm

Outline. 1 NP-Completeness Theory. 2 Limitation of Computation. 3 Examples. 4 Decision Problems. 5 Verification Algorithm Outline 1 NP-Completeness Theory 2 Limitation of Computation 3 Examples 4 Decision Problems 5 Verification Algorithm 6 Non-Deterministic Algorithm 7 NP-Complete Problems c Hu Ding (Michigan State University)

More information

Black Box Search By Unbiased Variation

Black Box Search By Unbiased Variation Black Box Search By Unbiased Variation Per Kristian Lehre and Carsten Witt CERCIA, University of Birmingham, UK DTU Informatics, Copenhagen, Denmark ThRaSH - March 24th 2010 State of the Art in Runtime

More information

Handling Uncertainty in Indicator-Based Multiobjective Optimization

Handling Uncertainty in Indicator-Based Multiobjective Optimization International Journal of Computational Intelligence Research. ISSN 0973-1873 Vol.2, No.3 (2006), pp. 255 272 c Research India Publications http://www.ijcir.info Handling Uncertainty in Indicator-Based

More information

COMP Analysis of Algorithms & Data Structures

COMP Analysis of Algorithms & Data Structures COMP 3170 - Analysis of Algorithms & Data Structures Shahin Kamali Computational Complexity CLRS 34.1-34.4 University of Manitoba COMP 3170 - Analysis of Algorithms & Data Structures 1 / 50 Polynomial

More information

Online Interval Coloring and Variants

Online Interval Coloring and Variants Online Interval Coloring and Variants Leah Epstein 1, and Meital Levy 1 Department of Mathematics, University of Haifa, 31905 Haifa, Israel. Email: lea@math.haifa.ac.il School of Computer Science, Tel-Aviv

More information

Covering Linear Orders with Posets

Covering Linear Orders with Posets Covering Linear Orders with Posets Proceso L. Fernandez, Lenwood S. Heath, Naren Ramakrishnan, and John Paul C. Vergara Department of Information Systems and Computer Science, Ateneo de Manila University,

More information

More on NP and Reductions

More on NP and Reductions Indian Institute of Information Technology Design and Manufacturing, Kancheepuram Chennai 600 127, India An Autonomous Institute under MHRD, Govt of India http://www.iiitdm.ac.in COM 501 Advanced Data

More information

Solving Zero-Sum Security Games in Discretized Spatio-Temporal Domains

Solving Zero-Sum Security Games in Discretized Spatio-Temporal Domains Solving Zero-Sum Security Games in Discretized Spatio-Temporal Domains APPENDIX LP Formulation for Constant Number of Resources (Fang et al. 3) For the sae of completeness, we describe the LP formulation

More information

P, NP, NP-Complete, and NPhard

P, NP, NP-Complete, and NPhard P, NP, NP-Complete, and NPhard Problems Zhenjiang Li 21/09/2011 Outline Algorithm time complicity P and NP problems NP-Complete and NP-Hard problems Algorithm time complicity Outline What is this course

More information

10.3 Matroids and approximation

10.3 Matroids and approximation 10.3 Matroids and approximation 137 10.3 Matroids and approximation Given a family F of subsets of some finite set X, called the ground-set, and a weight function assigning each element x X a non-negative

More information

Topics in Approximation Algorithms Solution for Homework 3

Topics in Approximation Algorithms Solution for Homework 3 Topics in Approximation Algorithms Solution for Homework 3 Problem 1 We show that any solution {U t } can be modified to satisfy U τ L τ as follows. Suppose U τ L τ, so there is a vertex v U τ but v L

More information

Computational Complexity and Genetic Algorithms

Computational Complexity and Genetic Algorithms Computational Complexity and Genetic Algorithms BART RYLANDER JAMES FOSTER School of Engineering Department of Computer Science University of Portland University of Idaho Portland, Or 97203 Moscow, Idaho

More information

U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 2 Luca Trevisan August 29, 2017

U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 2 Luca Trevisan August 29, 2017 U.C. Berkeley CS94: Beyond Worst-Case Analysis Handout Luca Trevisan August 9, 07 Scribe: Mahshid Montazer Lecture In this lecture, we study the Max Cut problem in random graphs. We compute the probable

More information

Connectedness of Efficient Solutions in Multiple. Objective Combinatorial Optimization

Connectedness of Efficient Solutions in Multiple. Objective Combinatorial Optimization Connectedness of Efficient Solutions in Multiple Objective Combinatorial Optimization Jochen Gorski Kathrin Klamroth Stefan Ruzika Communicated by H. Benson Abstract Connectedness of efficient solutions

More information

CSC 421: Algorithm Design & Analysis. Spring 2018

CSC 421: Algorithm Design & Analysis. Spring 2018 CSC 421: Algorithm Design & Analysis Spring 2018 Complexity & Computability complexity theory tractability, decidability P vs. NP, Turing machines NP-complete, reductions approximation algorithms, genetic

More information

Runtime Analysis of an Evolutionary Algorithm for Stochastic Multi-Objective Combinatorial Optimization

Runtime Analysis of an Evolutionary Algorithm for Stochastic Multi-Objective Combinatorial Optimization Runtime Analysis of an Evolutionary Algorithm for Stochastic Multi-Objective Combinatorial Optimization Walter J. Gutjahr Department of Statistics and Operations Research, University of Vienna, Universitaetsstr.

More information

Lecture 4 October 18th

Lecture 4 October 18th Directed and undirected graphical models Fall 2017 Lecture 4 October 18th Lecturer: Guillaume Obozinski Scribe: In this lecture, we will assume that all random variables are discrete, to keep notations

More information

Part V. Intractable Problems

Part V. Intractable Problems Part V Intractable Problems 507 Chapter 16 N P-Completeness Up to now, we have focused on developing efficient algorithms for solving problems. The word efficient is somewhat subjective, and the degree

More information

How to Treat Evolutionary Algorithms as Ordinary Randomized Algorithms

How to Treat Evolutionary Algorithms as Ordinary Randomized Algorithms How to Treat Evolutionary Algorithms as Ordinary Randomized Algorithms Technical University of Denmark (Workshop Computational Approaches to Evolution, March 19, 2014) 1/22 Context Running time/complexity

More information

Computability and Complexity Theory: An Introduction

Computability and Complexity Theory: An Introduction Computability and Complexity Theory: An Introduction meena@imsc.res.in http://www.imsc.res.in/ meena IMI-IISc, 20 July 2006 p. 1 Understanding Computation Kinds of questions we seek answers to: Is a given

More information

More Approximation Algorithms

More Approximation Algorithms CS 473: Algorithms, Spring 2018 More Approximation Algorithms Lecture 25 April 26, 2018 Most slides are courtesy Prof. Chekuri Ruta (UIUC) CS473 1 Spring 2018 1 / 28 Formal definition of approximation

More information

CS264: Beyond Worst-Case Analysis Lecture #11: LP Decoding

CS264: Beyond Worst-Case Analysis Lecture #11: LP Decoding CS264: Beyond Worst-Case Analysis Lecture #11: LP Decoding Tim Roughgarden October 29, 2014 1 Preamble This lecture covers our final subtopic within the exact and approximate recovery part of the course.

More information

Lecture 1 : Data Compression and Entropy

Lecture 1 : Data Compression and Entropy CPS290: Algorithmic Foundations of Data Science January 8, 207 Lecture : Data Compression and Entropy Lecturer: Kamesh Munagala Scribe: Kamesh Munagala In this lecture, we will study a simple model for

More information

Discrete Mathematics CS October 17, 2006

Discrete Mathematics CS October 17, 2006 Discrete Mathematics CS 2610 October 17, 2006 Uncountable sets Theorem: The set of real numbers is uncountable. If a subset of a set is uncountable, then the set is uncountable. The cardinality of a subset

More information

b + O(n d ) where a 1, b > 1, then O(n d log n) if a = b d d ) if a < b d O(n log b a ) if a > b d

b + O(n d ) where a 1, b > 1, then O(n d log n) if a = b d d ) if a < b d O(n log b a ) if a > b d CS161, Lecture 4 Median, Selection, and the Substitution Method Scribe: Albert Chen and Juliana Cook (2015), Sam Kim (2016), Gregory Valiant (2017) Date: January 23, 2017 1 Introduction Last lecture, we

More information

Algorithms and Data Structures 2016 Week 5 solutions (Tues 9th - Fri 12th February)

Algorithms and Data Structures 2016 Week 5 solutions (Tues 9th - Fri 12th February) Algorithms and Data Structures 016 Week 5 solutions (Tues 9th - Fri 1th February) 1. Draw the decision tree (under the assumption of all-distinct inputs) Quicksort for n = 3. answer: (of course you should

More information

(tree searching technique) (Boolean formulas) satisfying assignment: (X 1, X 2 )

(tree searching technique) (Boolean formulas) satisfying assignment: (X 1, X 2 ) Algorithms Chapter 5: The Tree Searching Strategy - Examples 1 / 11 Chapter 5: The Tree Searching Strategy 1. Ex 5.1Determine the satisfiability of the following Boolean formulas by depth-first search

More information

Behavior of EMO Algorithms on Many-Objective Optimization Problems with Correlated Objectives

Behavior of EMO Algorithms on Many-Objective Optimization Problems with Correlated Objectives H. Ishibuchi N. Akedo H. Ohyanagi and Y. Nojima Behavior of EMO algorithms on many-objective optimization problems with correlated objectives Proc. of 211 IEEE Congress on Evolutionary Computation pp.

More information

3.4 Relaxations and bounds

3.4 Relaxations and bounds 3.4 Relaxations and bounds Consider a generic Discrete Optimization problem z = min{c(x) : x X} with an optimal solution x X. In general, the algorithms generate not only a decreasing sequence of upper

More information

Geometric Steiner Trees

Geometric Steiner Trees Geometric Steiner Trees From the book: Optimal Interconnection Trees in the Plane By Marcus Brazil and Martin Zachariasen Part 3: Computational Complexity and the Steiner Tree Problem Marcus Brazil 2015

More information

Week 2: Sequences and Series

Week 2: Sequences and Series QF0: Quantitative Finance August 29, 207 Week 2: Sequences and Series Facilitator: Christopher Ting AY 207/208 Mathematicians have tried in vain to this day to discover some order in the sequence of prime

More information

PATH BUNDLES ON n-cubes

PATH BUNDLES ON n-cubes PATH BUNDLES ON n-cubes MATTHEW ELDER Abstract. A path bundle is a set of 2 a paths in an n-cube, denoted Q n, such that every path has the same length, the paths partition the vertices of Q n, the endpoints

More information

On Two Class-Constrained Versions of the Multiple Knapsack Problem

On Two Class-Constrained Versions of the Multiple Knapsack Problem On Two Class-Constrained Versions of the Multiple Knapsack Problem Hadas Shachnai Tami Tamir Department of Computer Science The Technion, Haifa 32000, Israel Abstract We study two variants of the classic

More information

Lecture 4: Proof of Shannon s theorem and an explicit code

Lecture 4: Proof of Shannon s theorem and an explicit code CSE 533: Error-Correcting Codes (Autumn 006 Lecture 4: Proof of Shannon s theorem and an explicit code October 11, 006 Lecturer: Venkatesan Guruswami Scribe: Atri Rudra 1 Overview Last lecture we stated

More information

Evolutionary Algorithms and Dynamic Programming

Evolutionary Algorithms and Dynamic Programming Evolutionary Algorithms and Dynamic Programming arxiv:1301.4096v1 [cs.ne] 17 Jan 2013 Benjamin Doerr, Anton Eremeev, Frank Neumann, Madeleine Theile, Christian Thyssen January 18, 2013 Abstract Recently,

More information

CSE 202 Homework 4 Matthias Springer, A

CSE 202 Homework 4 Matthias Springer, A CSE 202 Homework 4 Matthias Springer, A99500782 1 Problem 2 Basic Idea PERFECT ASSEMBLY N P: a permutation P of s i S is a certificate that can be checked in polynomial time by ensuring that P = S, and

More information

CS6999 Probabilistic Methods in Integer Programming Randomized Rounding Andrew D. Smith April 2003

CS6999 Probabilistic Methods in Integer Programming Randomized Rounding Andrew D. Smith April 2003 CS6999 Probabilistic Methods in Integer Programming Randomized Rounding April 2003 Overview 2 Background Randomized Rounding Handling Feasibility Derandomization Advanced Techniques Integer Programming

More information

1 Matchings in Non-Bipartite Graphs

1 Matchings in Non-Bipartite Graphs CS 598CSC: Combinatorial Optimization Lecture date: Feb 9, 010 Instructor: Chandra Chekuri Scribe: Matthew Yancey 1 Matchings in Non-Bipartite Graphs We discuss matching in general undirected graphs. Given

More information

On Constrained Boolean Pareto Optimization

On Constrained Boolean Pareto Optimization On Constrained Boolean Pareto Optimization Chao Qian and Yang Yu and Zhi-Hua Zhou National Key Laboratory for Novel Software Technology, Nanjing University Collaborative Innovation Center of Novel Software

More information

Switch Analysis for Running Time Analysis of Evolutionary Algorithms

Switch Analysis for Running Time Analysis of Evolutionary Algorithms IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. XX, NO. X, 204 Switch Analysis for Running Time Analysis of Evolutionary Algorithms Yang Yu, Member, IEEE, Chao Qian, Zhi-Hua Zhou, Fellow, IEEE Abstract

More information

Crossover Techniques in GAs

Crossover Techniques in GAs Crossover Techniques in GAs Debasis Samanta Indian Institute of Technology Kharagpur dsamanta@iitkgp.ac.in 16.03.2018 Debasis Samanta (IIT Kharagpur) Soft Computing Applications 16.03.2018 1 / 1 Important

More information

A An Overview of Complexity Theory for the Algorithm Designer

A An Overview of Complexity Theory for the Algorithm Designer A An Overview of Complexity Theory for the Algorithm Designer A.1 Certificates and the class NP A decision problem is one whose answer is either yes or no. Two examples are: SAT: Given a Boolean formula

More information

Combinatorial Optimization

Combinatorial Optimization Combinatorial Optimization 2017-2018 1 Maximum matching on bipartite graphs Given a graph G = (V, E), find a maximum cardinal matching. 1.1 Direct algorithms Theorem 1.1 (Petersen, 1891) A matching M is

More information

Lecture 2: Divide and conquer and Dynamic programming

Lecture 2: Divide and conquer and Dynamic programming Chapter 2 Lecture 2: Divide and conquer and Dynamic programming 2.1 Divide and Conquer Idea: - divide the problem into subproblems in linear time - solve subproblems recursively - combine the results in

More information

University of Toronto Department of Electrical and Computer Engineering. Final Examination. ECE 345 Algorithms and Data Structures Fall 2016

University of Toronto Department of Electrical and Computer Engineering. Final Examination. ECE 345 Algorithms and Data Structures Fall 2016 University of Toronto Department of Electrical and Computer Engineering Final Examination ECE 345 Algorithms and Data Structures Fall 2016 Print your first name, last name, UTORid, and student number neatly

More information

Multi-objective Quadratic Assignment Problem instances generator with a known optimum solution

Multi-objective Quadratic Assignment Problem instances generator with a known optimum solution Multi-objective Quadratic Assignment Problem instances generator with a known optimum solution Mădălina M. Drugan Artificial Intelligence lab, Vrije Universiteit Brussel, Pleinlaan 2, B-1050 Brussels,

More information

SMT 2013 Power Round Solutions February 2, 2013

SMT 2013 Power Round Solutions February 2, 2013 Introduction This Power Round is an exploration of numerical semigroups, mathematical structures which appear very naturally out of answers to simple questions. For example, suppose McDonald s sells Chicken

More information

Quick Sort Notes , Spring 2010

Quick Sort Notes , Spring 2010 Quick Sort Notes 18.310, Spring 2010 0.1 Randomized Median Finding In a previous lecture, we discussed the problem of finding the median of a list of m elements, or more generally the element of rank m.

More information

Lecture 1 : Probabilistic Method

Lecture 1 : Probabilistic Method IITM-CS6845: Theory Jan 04, 01 Lecturer: N.S.Narayanaswamy Lecture 1 : Probabilistic Method Scribe: R.Krithika The probabilistic method is a technique to deal with combinatorial problems by introducing

More information

Lecture 13 March 7, 2017

Lecture 13 March 7, 2017 CS 224: Advanced Algorithms Spring 2017 Prof. Jelani Nelson Lecture 13 March 7, 2017 Scribe: Hongyao Ma Today PTAS/FPTAS/FPRAS examples PTAS: knapsack FPTAS: knapsack FPRAS: DNF counting Approximation

More information

arxiv: v1 [math.co] 30 Nov 2016

arxiv: v1 [math.co] 30 Nov 2016 Rounds in a combinatorial search problem arxiv:1611.10133v1 [math.co] 30 Nov 2016 Dániel Gerbner Máté Vizer MTA Rényi Institute Hungary H-1053, Budapest, Reáltanoda utca 13-15. gerbner@renyi.hu, vizermate@gmail.com

More information

CS 374: Algorithms & Models of Computation, Spring 2017 Greedy Algorithms Lecture 19 April 4, 2017 Chandra Chekuri (UIUC) CS374 1 Spring / 1

CS 374: Algorithms & Models of Computation, Spring 2017 Greedy Algorithms Lecture 19 April 4, 2017 Chandra Chekuri (UIUC) CS374 1 Spring / 1 CS 374: Algorithms & Models of Computation, Spring 2017 Greedy Algorithms Lecture 19 April 4, 2017 Chandra Chekuri (UIUC) CS374 1 Spring 2017 1 / 1 Part I Greedy Algorithms: Tools and Techniques Chandra

More information

COS597D: Information Theory in Computer Science October 5, Lecture 6

COS597D: Information Theory in Computer Science October 5, Lecture 6 COS597D: Information Theory in Computer Science October 5, 2011 Lecture 6 Lecturer: Mark Braverman Scribe: Yonatan Naamad 1 A lower bound for perfect hash families. In the previous lecture, we saw that

More information

The Fitness Level Method with Tail Bounds

The Fitness Level Method with Tail Bounds The Fitness Level Method with Tail Bounds Carsten Witt DTU Compute Technical University of Denmark 2800 Kgs. Lyngby Denmark arxiv:307.4274v [cs.ne] 6 Jul 203 July 7, 203 Abstract The fitness-level method,

More information

The Knapsack Problem. 28. April /44

The Knapsack Problem. 28. April /44 The Knapsack Problem 20 10 15 20 W n items with weight w i N and profit p i N Choose a subset x of items Capacity constraint i x w i W wlog assume i w i > W, i : w i < W Maximize profit i x p i 28. April

More information

Week Cuts, Branch & Bound, and Lagrangean Relaxation

Week Cuts, Branch & Bound, and Lagrangean Relaxation Week 11 1 Integer Linear Programming This week we will discuss solution methods for solving integer linear programming problems. I will skip the part on complexity theory, Section 11.8, although this is

More information

Lecture 4: Two-point Sampling, Coupon Collector s problem

Lecture 4: Two-point Sampling, Coupon Collector s problem Randomized Algorithms Lecture 4: Two-point Sampling, Coupon Collector s problem Sotiris Nikoletseas Associate Professor CEID - ETY Course 2013-2014 Sotiris Nikoletseas, Associate Professor Randomized Algorithms

More information

Approximation Algorithms for Re-optimization

Approximation Algorithms for Re-optimization Approximation Algorithms for Re-optimization DRAFT PLEASE DO NOT CITE Dean Alderucci Table of Contents 1.Introduction... 2 2.Overview of the Current State of Re-Optimization Research... 3 2.1.General Results

More information

Combinatorial Optimization

Combinatorial Optimization Combinatorial Optimization Problem set 8: solutions 1. Fix constants a R and b > 1. For n N, let f(n) = n a and g(n) = b n. Prove that f(n) = o ( g(n) ). Solution. First we observe that g(n) 0 for all

More information

Tree sets. Reinhard Diestel

Tree sets. Reinhard Diestel 1 Tree sets Reinhard Diestel Abstract We study an abstract notion of tree structure which generalizes treedecompositions of graphs and matroids. Unlike tree-decompositions, which are too closely linked

More information

Crossover can be constructive when computing unique input output sequences

Crossover can be constructive when computing unique input output sequences Crossover can be constructive when computing unique input output sequences Per Kristian Lehre and Xin Yao The Centre of Excellence for Research in Computational Intelligence and Applications (CERCIA),

More information

A New Approach to Estimating the Expected First Hitting Time of Evolutionary Algorithms

A New Approach to Estimating the Expected First Hitting Time of Evolutionary Algorithms A New Approach to Estimating the Expected First Hitting Time of Evolutionary Algorithms Yang Yu and Zhi-Hua Zhou National Laboratory for Novel Software Technology Nanjing University, Nanjing 20093, China

More information

Weighted Activity Selection

Weighted Activity Selection Weighted Activity Selection Problem This problem is a generalization of the activity selection problem that we solvd with a greedy algorithm. Given a set of activities A = {[l, r ], [l, r ],..., [l n,

More information

Heuristics for The Whitehead Minimization Problem

Heuristics for The Whitehead Minimization Problem Heuristics for The Whitehead Minimization Problem R.M. Haralick, A.D. Miasnikov and A.G. Myasnikov November 11, 2004 Abstract In this paper we discuss several heuristic strategies which allow one to solve

More information

Efficient Approximation for Restricted Biclique Cover Problems

Efficient Approximation for Restricted Biclique Cover Problems algorithms Article Efficient Approximation for Restricted Biclique Cover Problems Alessandro Epasto 1, *, and Eli Upfal 2 ID 1 Google Research, New York, NY 10011, USA 2 Department of Computer Science,

More information

Basing Decisions on Sentences in Decision Diagrams

Basing Decisions on Sentences in Decision Diagrams Proceedings of the Twenty-Sixth AAAI Conference on Artificial Intelligence Basing Decisions on Sentences in Decision Diagrams Yexiang Xue Department of Computer Science Cornell University yexiang@cs.cornell.edu

More information