1 The Knapsack Problem
|
|
- Wilfred Lynch
- 5 years ago
- Views:
Transcription
1 Comp 260: Advanced Algorithms Prof. Lenore Cowen Tufts University, Spring 2018 Scribe: Tom Magerlein 1 Lecture 4: The Knapsack Problem 1 The Knapsack Problem Suppose we are trying to burgle someone s house. In this house, we find a collection of objects S = a 1, a 2,..., a n 1, a n s 1, s 2,..., s n 1, s n p 1, p 2,..., p n 1, p n the objects names their sizes their profits We have a knapsack with capacity B, into which we would like to put these objects. Our goal is to find a subset of these objects whose total size is no larger than B, and whose total profit is maximized. We can view this in terms of indicator variables, where { 1 if a i is chosen x i = 0 otherwise We then need to find a setting of these x i s to maximize x i p i subject to the constraint that x i s i B. This would be a very simple linear program, but we have an integer constraint, since no fence will pay you one third of the price if you bring them one third of a piccolo. Thus, the knapsack problem is an integer program that is NP-Hard. We would like to find a polynomial time approximation to the optimal solution, so that we can leave the house we are burgling before the residents get home. 1 Based on previous version scribed by Kristen McNeely-Shaw 1
2 2 Running example We will refer to this example throughout the remainder of the notes. Object A B C D E Size Profit Knapsack size: B 3 A First Guess at Approximation For each i, we consider the profit-to-size ratio p i, and reassign the objects s i indices such that p 1 p 2 p n s 1 s 2 s n Referring back to the running example, this yields Object A B C D E Profit-Size Ratio 3/7 1 1/3 1/3 2 Order a 3 a 2 a 4 a 5 a 1 Intuitively, this organizes the objects by profit density, or profit per unit size. Algorithm 1 (Greedy): pick the first k objects greedily in this order until the next object, a k+1, will not fit in the knapsack. We can construct a simple example to show that this will not always choose the optimal collection of objects: Chosen by Algorithm 1 ((100 B) 1) B More profitable overall 2
3 Algorithm 1 chooses the first object, because it has a slightly higher profitto-size ratio, but choosing it leaves much of the knapsack empty. Note that by setting the profit appropriately, we can generalize this example to make the approximation ratio arbitrarily bad. 4 Improving the Algorithm We can modify the greedy algorithm to compensate for this issue: Algorithm 2 (Smart Greedy): pick the more profitable of {a 1,..., a k } and {a k+1 }. This algorithm is still not optimal. Consider the example below: 5/50 5/51 9/100 Chosen by Algorithm 1 Chosen by Algorithm 2 Most profitable overall However, we are now going to be able to prove a bound on the approximation ratio. Claim 4.1. Algorithm 2 always gives a profit at least 0.5 OP T, i.e. if the profit of the optimal solution is P, then the profit of the solution found by Algorithm 2 is at least P 2. If we fill a knapsack to capacity, the most profit we can get is in order of profit density: {a 1 a k } a k+1 Figure 1: A knapsack filled in order of profit ratio. Thus if we increased the size of the knapsack just enough to also allow space for the next most profit-dense object, this would be an optimal profit for the 3
4 slightly bigger knapsack! So it s certainly an upper bound on the profit of the original knapsack, and thus: Lemma 4.2. The optimal profit P p 1 + p p k + p k+1. From this lemma, it follows that ( P 2 max p k+1, ) k p i i=1 5 How close can we get to OPT? We now show that no algorithm gets within an additive factor of OPT unless P = NP. In the next section, we then show that we can get arbitrarily close to optimal by a multiplicative factor. Theorem 5.1. If P N P, no polynomial time approximation algorithm can solve knapsack with value P k for any fixed constant k. Proof. By contradiction. Assume there exists a polynomial time algorithm A with performance guarantee k (an integer) for all instances of knapsack. That is, for any knapsack instance I, A(I) = P k. We show that A can be used to construct a solution to knapsack with value P in polynomial time. Suppose we have an instance of knapsack: I = { a i, p i, s i } for i = 1... n Let I = { a i, p i, s i } where a i = a i, p 1 = p i (k + 1) and s i = s i. Note: a feasible solution for I corresponds to a feasible solution for I, and the value of a solution for I is exactly (k+1) times the value of the corresponding solution for I. 4
5 Let M = A(I). If we run A on I, the solution satisfies the following: Since A(I ) P (I ) k M (k + 1) P (I) (k + 1) k M P (I) k k + 1 k k + 1 < 1 and both M and P (I) must be integers, this gives that M P (I) 0 This in turn implies that M is an optimal solution for I, and thus that A is a polynomial time algorithm to determine an exact solution to knapsack, which we above demonstrated was impossible. Thus, no such algorithm A exists. 6 Approximation Schemes Definition 6.1. Let π be an optimization problem with objective function f π and optimal solution S. We say that A is an approximation scheme for π if on input (I, ɛ) where I is an instance of π and ɛ > 0 is a fixed error parameter, it outputs a solution S such that f π (I, S) (1 + ɛ) S if π is a minimization problem f π (I, S) (1 ɛ) S if π is a maximization problem Definition 6.2. A is a Polynomial Time Approximation Scheme (PTAS) if it is an approximation scheme where for each fixed ɛ > 0, its running time is polynomial in the size of instance I. Note that an PTAS is polynomial time in the size of I its time complexity can depend arbitrarily badly on ɛ, for example n /ɛ or n 3 21/ɛ 5
6 Definition 6.3. A is an Fully Polynomial Time Approximation Scheme (FPTAS) if it is a PTAS, and its running time is also polynomial in 1/ɛ. Remark: If P NP, an FPTAS is the best you can do for an NP-Hard problem. 7 An FPTAS for Knapsack using Dynamic Programming We now show that knapsack has an FPTAS, meaning that, at least in the approximate sense, it is an easy NP-hard problem. Recall our running example Objects A B C D E Sizes Profits Knapsack size: B Let P max be the profit of the most profitable object i.e. P max = max i S (p i ). Trivially, n P max P. Use the original arbitrary ordering. For each count i = 1,..., n and profit p = 1,..., n P max, let S i,p denote a subset of objects a 1,..., a i whose total profit is exactly p, and whose total space is minimized. Let A(i, p) denote the space that set S i,p fills if it exists, or if no such set exists. It is clear that P = max(p A(n, p) B) We now show how to compute A(i, p) for all i = 1,..., n and p = 1,..., n P max in time O(n 2 P max ) using dynamic programming. Note: This is not a contradiction to the NP-hardness of knapsack, because P max isn t necessarily bounded by a polynomial in n (although it would be if we required profits to be represented in unary). 6
7 All of the values of A(i, p) are computed as follows: A(1, p) is known for every value of p, so begin by initializing the table so that A(1, p 1 ) = s 1 and A(1, p i 1 ) =. Count i Profit p Further cells in the table can be filled in using the recurrence relation A(i + 1, p) = min{a(i, p), A(i, p p i+1 ) + s i+1 } and treating A(i, 0) = 0 for all i. This recurrence relation comes from the fact that, given the lowest-space ways to attain each value of profit using the first k items, there are two possibilities for the minimum-space way to attain a particular profit when the next item is added. First, it may be that the new item does not allow a lower-space way of reaching the desired profit; in this case, the new item is not taken, so the optimal collection of items is the same as in the previous row of the table. Alternatively, the new item may be taken, along with the lowest-space collection of additional items needed to reach the appropriate profit; we already know what this collection is since we have filled in the entire previous line of the table. The recurrence relation then just takes whichever of these two uses the least total space. Evaluating the relation is constant time in each cell, and thus it costs O(n 2 P max ) to fill the table. Count i Profit p
8 8 What to Do When Profits are Too Large When P max is bounded by a polynomial in the size of the input, then the above is a polynomial time algorithm for knapsack. When it is not, we now present a (1 ɛ) approximation algorithm, by rounding the profit values. Algorithm 3 (FPTAS): 1. Given an ɛ > 0, let k = ɛ P max. n 2. For each a i, define p pi i =. k 3. Let I = ( a i, s i, p i ) where a i = a i, s i = s i, and p i as above. 4. Using the dynamic programming method from the previous section, find the most profitable set S. 5. Output max(s max, S ), where S max is the smallest object of profit P max if S max B, and otherwise is 0. Lemma 8.1. Let A denote the set output by Algorithm 3. Then Profit(A) (1 ɛ) P Proof. Let O denote the set with profit P. For any a i, k p i can be smaller than p i, but not by more than k. Thus, Profit(O) k Profit (O) n k Under the Profit scheme, Profit (S ) is OPTIMAL, i.e. Profit (S ) Profit (Y ) for all Y, and in particular Profit (S ) Profit (O). Thus: Profit(S ) k Profit (S ) k Profit (O) Profit(O) n k = P ɛ P max 8
9 Since the algorithm also considers the most profitable element Profit(A) Profit(P max ) Profit(A) Profit(S ) P ɛ Profit(A)s Theorem 8.2. Algorithm 3 is an FPTAS for knapsack. Proof. By lemma, the solution P is within (1 ɛ) of OPT. ( ) The running time is O n 2 Pmax ( n ) = O n 2, which is polynomial in k ɛ n and 1/ɛ. 9 Strong NP-Hardness Many NP-hard problems are more difficult to approximate than knapsack, and do not admit an FPTAS: Definition 9.1. A problem π is strongly NP-hard if every problem in NP can be polynomial-time reduced to π so that all numbers in the reduced instance can be written in unary. Theorem 9.2. Let p be a polynomial and π be a strongly NP-hard minimization problem such that the objective function f π is integer valued, and on any instance I, the size of OP T (I) < p( I ). Then π does not admit an FPTAS, assuming P NP. We don t prove this theorem here, although the proof is not that difficult. Note that this implies that knapsack is not strongly NP-hard, as we above constructed an FPTAS to approximate it. 9
Lecture 4: An FPTAS for Knapsack, and K-Center
Comp 260: Advanced Algorithms Tufts University, Spring 2016 Prof. Lenore Cowen Scribe: Eric Bailey Lecture 4: An FPTAS for Knapsack, and K-Center 1 Introduction Definition 1.0.1. The Knapsack problem (restated)
More information8 Knapsack Problem 8.1 (Knapsack)
8 Knapsack In Chapter 1 we mentioned that some NP-hard optimization problems allow approximability to any required degree. In this chapter, we will formalize this notion and will show that the knapsack
More informationLecture 11 October 7, 2013
CS 4: Advanced Algorithms Fall 03 Prof. Jelani Nelson Lecture October 7, 03 Scribe: David Ding Overview In the last lecture we talked about set cover: Sets S,..., S m {,..., n}. S has cost c S. Goal: Cover
More informationThis means that we can assume each list ) is
This means that we can assume each list ) is of the form ),, ( )with < and Since the sizes of the items are integers, there are at most +1pairs in each list Furthermore, if we let = be the maximum possible
More informationKnapsack. Bag/knapsack of integer capacity B n items item i has size s i and profit/weight w i
Knapsack Bag/knapsack of integer capacity B n items item i has size s i and profit/weight w i Goal: find a subset of items of maximum profit such that the item subset fits in the bag Knapsack X: item set
More informationa 1 a 2 a 3 a 4 v i c i c(a 1, a 3 ) = 3
AM 221: Advanced Optimization Spring 2016 Prof. Yaron Singer Lecture 17 March 30th 1 Overview In the previous lecture, we saw examples of combinatorial problems: the Maximal Matching problem and the Minimum
More informationLecture 13 March 7, 2017
CS 224: Advanced Algorithms Spring 2017 Prof. Jelani Nelson Lecture 13 March 7, 2017 Scribe: Hongyao Ma Today PTAS/FPTAS/FPRAS examples PTAS: knapsack FPTAS: knapsack FPRAS: DNF counting Approximation
More informationNETS 412: Algorithmic Game Theory March 28 and 30, Lecture Approximation in Mechanism Design. X(v) = arg max v i (a)
NETS 412: Algorithmic Game Theory March 28 and 30, 2017 Lecture 16+17 Lecturer: Aaron Roth Scribe: Aaron Roth Approximation in Mechanism Design In the last lecture, we asked how far we can go beyond the
More informationNetwork Design and Game Theory Spring 2008 Lecture 2
Network Design and Game Theory Spring 2008 Lecture 2 Instructor: Mohammad T. Hajiaghayi Scribe: Imdadullah Khan February 04, 2008 MAXIMUM COVERAGE In this lecture we review Maximum Coverage and Unique
More informationCOSC 341: Lecture 25 Coping with NP-hardness (2)
1 Introduction Figure 1: Famous cartoon by Garey and Johnson, 1979 We have seen the definition of a constant factor approximation algorithm. The following is something even better. 2 Approximation Schemes
More informationChapter 11. Approximation Algorithms. Slides by Kevin Wayne Pearson-Addison Wesley. All rights reserved.
Chapter 11 Approximation Algorithms Slides by Kevin Wayne. Copyright @ 2005 Pearson-Addison Wesley. All rights reserved. 1 Approximation Algorithms Q. Suppose I need to solve an NP-hard problem. What should
More information1 Review for Lecture 2 MaxFlow
Comp 260: Advanced Algorithms Tufts University, Spring 2009 Prof. Lenore Cowen Scribe: Wanyu Wang Lecture 13: Back to MaxFlow/Edmonds-Karp 1 Review for Lecture 2 MaxFlow A flow network showing flow and
More informationThe Knapsack Problem. 28. April /44
The Knapsack Problem 20 10 15 20 W n items with weight w i N and profit p i N Choose a subset x of items Capacity constraint i x w i W wlog assume i w i > W, i : w i < W Maximize profit i x p i 28. April
More informationLec. 2: Approximation Algorithms for NP-hard Problems (Part II)
Limits of Approximation Algorithms 28 Jan, 2010 (TIFR) Lec. 2: Approximation Algorithms for NP-hard Problems (Part II) Lecturer: Prahladh Harsha Scribe: S. Ajesh Babu We will continue the survey of approximation
More informationBin packing and scheduling
Sanders/van Stee: Approximations- und Online-Algorithmen 1 Bin packing and scheduling Overview Bin packing: problem definition Simple 2-approximation (Next Fit) Better than 3/2 is not possible Asymptotic
More information1 Ordinary Load Balancing
Comp 260: Advanced Algorithms Prof. Lenore Cowen Tufts University, Spring 208 Scribe: Emily Davis Lecture 8: Scheduling Ordinary Load Balancing Suppose we have a set of jobs each with their own finite
More informationMore Approximation Algorithms
CS 473: Algorithms, Spring 2018 More Approximation Algorithms Lecture 25 April 26, 2018 Most slides are courtesy Prof. Chekuri Ruta (UIUC) CS473 1 Spring 2018 1 / 28 Formal definition of approximation
More information0-1 Knapsack Problem
KP-0 0-1 Knapsack Problem Define object o i with profit p i > 0 and weight w i > 0, for 1 i n. Given n objects and a knapsack capacity C > 0, the problem is to select a subset of objects with largest total
More information1 Column Generation and the Cutting Stock Problem
1 Column Generation and the Cutting Stock Problem In the linear programming approach to the traveling salesman problem we used the cutting plane approach. The cutting plane approach is appropriate when
More informationLecture 4. 1 FPTAS - Fully Polynomial Time Approximation Scheme
Theory of Computer Science to Msc Students, Spring 2007 Lecturer: Dorit Aharonov Lecture 4 Scribe: Ram Bouobza & Yair Yarom Revised: Shahar Dobzinsi, March 2007 1 FPTAS - Fully Polynomial Time Approximation
More informationChapter 11. Approximation Algorithms. Slides by Kevin Wayne Pearson-Addison Wesley. All rights reserved.
Chapter 11 Approximation Algorithms Slides by Kevin Wayne. Copyright @ 2005 Pearson-Addison Wesley. All rights reserved. 1 P and NP P: The family of problems that can be solved quickly in polynomial time.
More informationAlgorithms. Outline! Approximation Algorithms. The class APX. The intelligence behind the hardware. ! Based on
6117CIT - Adv Topics in Computing Sci at Nathan 1 Algorithms The intelligence behind the hardware Outline! Approximation Algorithms The class APX! Some complexity classes, like PTAS and FPTAS! Illustration
More informationLecture 20: LP Relaxation and Approximation Algorithms. 1 Introduction. 2 Vertex Cover problem. CSCI-B609: A Theorist s Toolkit, Fall 2016 Nov 8
CSCI-B609: A Theorist s Toolkit, Fall 2016 Nov 8 Lecture 20: LP Relaxation and Approximation Algorithms Lecturer: Yuan Zhou Scribe: Syed Mahbub Hafiz 1 Introduction When variables of constraints of an
More informationAside: Golden Ratio. Golden Ratio: A universal law. Golden ratio φ = lim n = 1+ b n = a n 1. a n+1 = a n + b n, a n+b n a n
Aside: Golden Ratio Golden Ratio: A universal law. Golden ratio φ = lim n a n+b n a n = 1+ 5 2 a n+1 = a n + b n, b n = a n 1 Ruta (UIUC) CS473 1 Spring 2018 1 / 41 CS 473: Algorithms, Spring 2018 Dynamic
More informationP,NP, NP-Hard and NP-Complete
P,NP, NP-Hard and NP-Complete We can categorize the problem space into two parts Solvable Problems Unsolvable problems 7/11/2011 1 Halting Problem Given a description of a program and a finite input, decide
More informationA Fast Asymptotic Approximation Scheme for Bin Packing with Rejection
A Fast Asymptotic Approximation Scheme for Bin Packing with Rejection p. 1/3 A Fast Asymptotic Approximation Scheme for Bin Packing with Rejection Wolfgang Bein 1, José R. Correa 2, Xin Han 3 1 School
More information1 Maximum Budgeted Allocation
CS 369P: Polyhedral techniques in combinatorial optimization Instructor: Jan Vondrák Lecture date: November 4, 2010 Scribe: David Tobin 1 Maximum Budgeted Allocation Agents Items Given: n agents and m
More informationApproximation Algorithms
Approximation Algorithms What do you do when a problem is NP-complete? or, when the polynomial time solution is impractically slow? assume input is random, do expected performance. Eg, Hamiltonian path
More informationFundamentals of optimization problems
Fundamentals of optimization problems Dmitriy Serdyuk Ferienakademie in Sarntal 2012 FAU Erlangen-Nürnberg, TU München, Uni Stuttgart September 2012 Overview 1 Introduction Optimization problems PO and
More informationarxiv: v1 [math.oc] 3 Jan 2019
The Product Knapsack Problem: Approximation and Complexity arxiv:1901.00695v1 [math.oc] 3 Jan 2019 Ulrich Pferschy a, Joachim Schauer a, Clemens Thielen b a Department of Statistics and Operations Research,
More informationIntractable Problems Part Two
Intractable Problems Part Two Announcements Problem Set Five graded; will be returned at the end of lecture. Extra office hours today after lecture from 4PM 6PM in Clark S250. Reminder: Final project goes
More informationCSE 421 Dynamic Programming
CSE Dynamic Programming Yin Tat Lee Weighted Interval Scheduling Interval Scheduling Job j starts at s(j) and finishes at f j and has weight w j Two jobs compatible if they don t overlap. Goal: find maximum
More informationLecture 13: Dynamic Programming Part 2 10:00 AM, Feb 23, 2018
CS18 Integrated Introduction to Computer Science Fisler, Nelson Lecture 13: Dynamic Programming Part 2 10:00 AM, Feb 23, 2018 Contents 1 Holidays 1 1.1 Halloween..........................................
More informationBCCS: Computational methods for complex systems Wednesday, 16 June Lecture 3
BCCS: Computational methods for complex systems Wednesday, 16 June 2010 Lecturer: Ashley Montanaro 1 Lecture 3 Hardness of approximation 1 Overview In my lectures so far, we ve mostly restricted ourselves
More information1 T 1 = where 1 is the all-ones vector. For the upper bound, let v 1 be the eigenvector corresponding. u:(u,v) E v 1(u)
CME 305: Discrete Mathematics and Algorithms Instructor: Reza Zadeh (rezab@stanford.edu) Final Review Session 03/20/17 1. Let G = (V, E) be an unweighted, undirected graph. Let λ 1 be the maximum eigenvalue
More information1 Agenda. 2 History. 3 Probabilistically Checkable Proofs (PCPs). Lecture Notes Definitions. PCPs. Approximation Algorithms.
CS 221: Computational Complexity Prof. Salil Vadhan Lecture Notes 20 April 12, 2010 Scribe: Jonathan Pines 1 Agenda. PCPs. Approximation Algorithms. PCPs = Inapproximability. 2 History. First, some history
More informationOn Two Class-Constrained Versions of the Multiple Knapsack Problem
On Two Class-Constrained Versions of the Multiple Knapsack Problem Hadas Shachnai Tami Tamir Department of Computer Science The Technion, Haifa 32000, Israel Abstract We study two variants of the classic
More informationOptimization Methods in Management Science
Optimization Methods in Management Science MIT 15.05 Recitation 8 TAs: Giacomo Nannicini, Ebrahim Nasrabadi At the end of this recitation, students should be able to: 1. Derive Gomory cut from fractional
More informationOn the Tightness of an LP Relaxation for Rational Optimization and its Applications
OPERATIONS RESEARCH Vol. 00, No. 0, Xxxxx 0000, pp. 000 000 issn 0030-364X eissn 526-5463 00 0000 000 INFORMS doi 0.287/xxxx.0000.0000 c 0000 INFORMS Authors are encouraged to submit new papers to INFORMS
More informationApproximation Algorithms for the Incremental Knapsack Problem via Disjunctive Programming
Approximation Algorithms for the Incremental Knapsack Problem via Disjunctive Programming Daniel Bienstock, Jay Sethuraman, Chun Ye Department of Industrial Engineering and Operations Research Columbia
More informationSection Notes 9. IP: Cutting Planes. Applied Math 121. Week of April 12, 2010
Section Notes 9 IP: Cutting Planes Applied Math 121 Week of April 12, 2010 Goals for the week understand what a strong formulations is. be familiar with the cutting planes algorithm and the types of cuts
More informationThe Knapsack Problem. n items with weight w i N and profit p i N. Choose a subset x of items
Sanders/van Stee: Approximations- und Online-Algorithmen 1 The Knapsack Problem 10 15 W n items with weight w i N and profit p i N Choose a subset x of items Capacity constraint i x w i W wlog assume i
More informationCOL351: Analysis and Design of Algorithms (CSE, IITD, Semester-I ) Name: Entry number:
Name: Entry number: There are 5 questions for a total of 75 points. 1. (5 points) You are given n items and a sack that can hold at most W units of weight. The weight of the i th item is denoted by w(i)
More informationA Polynomial Time Approximation Scheme for the Multiple Knapsack Problem
University of Pennsylvania ScholarlyCommons Departmental Papers (CIS) Department of Computer & Information Science February 2006 A Polynomial Time Approximation Scheme for the Multiple Knapsack Problem
More informationDual fitting approximation for Set Cover, and Primal Dual approximation for Set Cover
duality 1 Dual fitting approximation for Set Cover, and Primal Dual approximation for Set Cover Guy Kortsarz duality 2 The set cover problem with uniform costs Input: A universe U and a collection of subsets
More informationOptimisation and Operations Research
Optimisation and Operations Research Lecture 15: The Greedy Heuristic Matthew Roughan http://www.maths.adelaide.edu.au/matthew.roughan/ Lecture_notes/OORII/ School of
More informationImproved Fully Polynomial time Approximation Scheme for the 0-1 Multiple-choice Knapsack Problem
Improved Fully Polynomial time Approximation Scheme for the 0-1 Multiple-choice Knapsack Problem Mukul Subodh Bansal V.Ch.Venkaiah International Institute of Information Technology Hyderabad, India Abstract
More informationCS264: Beyond Worst-Case Analysis Lecture #11: LP Decoding
CS264: Beyond Worst-Case Analysis Lecture #11: LP Decoding Tim Roughgarden October 29, 2014 1 Preamble This lecture covers our final subtopic within the exact and approximate recovery part of the course.
More informationInteger Linear Programs
Lecture 2: Review, Linear Programming Relaxations Today we will talk about expressing combinatorial problems as mathematical programs, specifically Integer Linear Programs (ILPs). We then see what happens
More informationRevisiting the Greedy Approach to Submodular Set Function Maximization
Submitted to manuscript Revisiting the Greedy Approach to Submodular Set Function Maximization Pranava R. Goundan Analytics Operations Engineering, pranava@alum.mit.edu Andreas S. Schulz MIT Sloan School
More information10.3 Matroids and approximation
10.3 Matroids and approximation 137 10.3 Matroids and approximation Given a family F of subsets of some finite set X, called the ground-set, and a weight function assigning each element x X a non-negative
More informationSanta Claus Schedules Jobs on Unrelated Machines
Santa Claus Schedules Jobs on Unrelated Machines Ola Svensson (osven@kth.se) Royal Institute of Technology - KTH Stockholm, Sweden March 22, 2011 arxiv:1011.1168v2 [cs.ds] 21 Mar 2011 Abstract One of the
More informationAsymptotic Polynomial-Time Approximation (APTAS) and Randomized Approximation Algorithms
Approximation Algorithms Asymptotic Polynomial-Time Approximation (APTAS) and Randomized Approximation Algorithms Jens Egeblad November 29th, 2006 Agenda First lesson (Asymptotic Approximation): Bin-Packing
More informationDynamic Programming: Interval Scheduling and Knapsack
Dynamic Programming: Interval Scheduling and Knapsack . Weighted Interval Scheduling Weighted Interval Scheduling Weighted interval scheduling problem. Job j starts at s j, finishes at f j, and has weight
More informationLecture 6,7 (Sept 27 and 29, 2011 ): Bin Packing, MAX-SAT
,7 CMPUT 675: Approximation Algorithms Fall 2011 Lecture 6,7 (Sept 27 and 29, 2011 ): Bin Pacing, MAX-SAT Lecturer: Mohammad R. Salavatipour Scribe: Weitian Tong 6.1 Bin Pacing Problem Recall the bin pacing
More informationLecturer: Shuchi Chawla Topic: Inapproximability Date: 4/27/2007
CS880: Approximations Algorithms Scribe: Tom Watson Lecturer: Shuchi Chawla Topic: Inapproximability Date: 4/27/2007 So far in this course, we have been proving upper bounds on the approximation factors
More informationTopic: Intro, Vertex Cover, TSP, Steiner Tree Date: 1/23/2007
CS880: Approximations Algorithms Scribe: Michael Kowalczyk Lecturer: Shuchi Chawla Topic: Intro, Vertex Cover, TSP, Steiner Tree Date: 1/23/2007 Today we discuss the background and motivation behind studying
More information9. Submodular function optimization
Submodular function maximization 9-9. Submodular function optimization Submodular function maximization Greedy algorithm for monotone case Influence maximization Greedy algorithm for non-monotone case
More informationCS 583: Approximation Algorithms: Introduction
CS 583: Approximation Algorithms: Introduction Chandra Chekuri January 15, 2018 1 Introduction Course Objectives 1. To appreciate that not all intractable problems are the same. NP optimization problems,
More informationIntroduction to Analysis Constructing R from Q
Introduction to Analysis Constructing R from Q Definition. A subset A Q is called a cut if it posses the following three properties. A and A Q. 2. If r A, then also A contains every rational q < r. 3.
More informationCSCI3390-Lecture 17: A sampler of NP-complete problems
CSCI3390-Lecture 17: A sampler of NP-complete problems 1 List of Problems We now know that if L is any problem in NP, that L P SAT, and thus SAT is NP-hard. Since SAT is also in NP we find that SAT is
More informationThe knapsack problem
Moderne Suchmethoden der Informatik The knapsack problem Florian Pawlik December 1, 2014 Abstract Have you ever wondered about how heavy one can fill his knapsack? No? That does not matter You will get
More informationCS 6901 (Applied Algorithms) Lecture 3
CS 6901 (Applied Algorithms) Lecture 3 Antonina Kolokolova September 16, 2014 1 Representative problems: brief overview In this lecture we will look at several problems which, although look somewhat similar
More informationLectures 6, 7 and part of 8
Lectures 6, 7 and part of 8 Uriel Feige April 26, May 3, May 10, 2015 1 Linear programming duality 1.1 The diet problem revisited Recall the diet problem from Lecture 1. There are n foods, m nutrients,
More informationLecture 10 September 27, 2016
CS 395T: Sublinear Algorithms Fall 2016 Prof. Eric Price Lecture 10 September 27, 2016 Scribes: Quinten McNamara & William Hoza 1 Overview In this lecture, we focus on constructing coresets, which are
More informationEssential facts about NP-completeness:
CMPSCI611: NP Completeness Lecture 17 Essential facts about NP-completeness: Any NP-complete problem can be solved by a simple, but exponentially slow algorithm. We don t have polynomial-time solutions
More informationA note on semi-online machine covering
A note on semi-online machine covering Tomáš Ebenlendr 1, John Noga 2, Jiří Sgall 1, and Gerhard Woeginger 3 1 Mathematical Institute, AS CR, Žitná 25, CZ-11567 Praha 1, The Czech Republic. Email: ebik,sgall@math.cas.cz.
More informationMechanisms for Multi-Unit Auctions
Journal of Artificial Intelligence Research 37 (2010) 85-98 Submitted 10/09; published 2/10 Mechanisms for Multi-Unit Auctions Shahar Dobzinski Computer Science Department, Cornell University Ithaca, NY
More informationApproximation Algorithms (Load Balancing)
July 6, 204 Problem Definition : We are given a set of n jobs {J, J 2,..., J n }. Each job J i has a processing time t i 0. We are given m identical machines. Problem Definition : We are given a set of
More informationFinding Large Induced Subgraphs
Finding Large Induced Subgraphs Illinois Institute of Technology www.math.iit.edu/ kaul kaul@math.iit.edu Joint work with S. Kapoor, M. Pelsmajer (IIT) Largest Induced Subgraph Given a graph G, find the
More informationc i r i i=1 r 1 = [1, 2] r 2 = [0, 1] r 3 = [3, 4].
Lecture Notes: Rank of a Matrix Yufei Tao Department of Computer Science and Engineering Chinese University of Hong Kong taoyf@cse.cuhk.edu.hk 1 Linear Independence Definition 1. Let r 1, r 2,..., r m
More information1 The Arthur-Merlin Story
Comp 260: Advanced Algorithms Tufts University, Spring 2018 Prof. Lenore Cowen Scribe: Elias Marcopoulos Lecture 1a: Bipartite Perfect Matching 1 The Arthur-Merlin Story In the land ruled by the legendary
More informationApproximation Basics
Approximation Basics, Concepts, and Examples Xiaofeng Gao Department of Computer Science and Engineering Shanghai Jiao Tong University, P.R.China Fall 2012 Special thanks is given to Dr. Guoqiang Li for
More informationLecture 6: Greedy Algorithms I
COMPSCI 330: Design and Analysis of Algorithms September 14 Lecturer: Rong Ge Lecture 6: Greedy Algorithms I Scribe: Fred Zhang 1 Overview In this lecture, we introduce a new algorithm design technique
More informationNP Completeness and Approximation Algorithms
Chapter 10 NP Completeness and Approximation Algorithms Let C() be a class of problems defined by some property. We are interested in characterizing the hardest problems in the class, so that if we can
More informationCS 6783 (Applied Algorithms) Lecture 3
CS 6783 (Applied Algorithms) Lecture 3 Antonina Kolokolova January 14, 2013 1 Representative problems: brief overview of the course In this lecture we will look at several problems which, although look
More informationWeek Cuts, Branch & Bound, and Lagrangean Relaxation
Week 11 1 Integer Linear Programming This week we will discuss solution methods for solving integer linear programming problems. I will skip the part on complexity theory, Section 11.8, although this is
More informationOutline. 1 NP-Completeness Theory. 2 Limitation of Computation. 3 Examples. 4 Decision Problems. 5 Verification Algorithm
Outline 1 NP-Completeness Theory 2 Limitation of Computation 3 Examples 4 Decision Problems 5 Verification Algorithm 6 Non-Deterministic Algorithm 7 NP-Complete Problems c Hu Ding (Michigan State University)
More informationLecture 3. 1 Polynomial-time algorithms for the maximum flow problem
ORIE 633 Network Flows August 30, 2007 Lecturer: David P. Williamson Lecture 3 Scribe: Gema Plaza-Martínez 1 Polynomial-time algorithms for the maximum flow problem 1.1 Introduction Let s turn now to considering
More information(tree searching technique) (Boolean formulas) satisfying assignment: (X 1, X 2 )
Algorithms Chapter 5: The Tree Searching Strategy - Examples 1 / 11 Chapter 5: The Tree Searching Strategy 1. Ex 5.1Determine the satisfiability of the following Boolean formulas by depth-first search
More informationApproximation Algorithms for Orthogonal Packing Problems for Hypercubes
Approximation Algorithms for Orthogonal Packing Problems for Hypercubes Rolf Harren 1 Max-Planck-Institut für Informatik, Campus E 1 4, 66123 Saarbrücken, Germany Abstract Orthogonal packing problems are
More informationOptimization of Submodular Functions Tutorial - lecture I
Optimization of Submodular Functions Tutorial - lecture I Jan Vondrák 1 1 IBM Almaden Research Center San Jose, CA Jan Vondrák (IBM Almaden) Submodular Optimization Tutorial 1 / 1 Lecture I: outline 1
More informationGame Theory. Greg Plaxton Theory in Programming Practice, Spring 2004 Department of Computer Science University of Texas at Austin
Game Theory Greg Plaxton Theory in Programming Practice, Spring 2004 Department of Computer Science University of Texas at Austin Bimatrix Games We are given two real m n matrices A = (a ij ), B = (b ij
More informationMat 3770 Bin Packing or
Basic Algithm Spring 2014 Used when a problem can be partitioned into non independent sub problems Basic Algithm Solve each sub problem once; solution is saved f use in other sub problems Combine solutions
More informationNP-COMPLETE PROBLEMS. 1. Characterizing NP. Proof
T-79.5103 / Autumn 2006 NP-complete problems 1 NP-COMPLETE PROBLEMS Characterizing NP Variants of satisfiability Graph-theoretic problems Coloring problems Sets and numbers Pseudopolynomial algorithms
More informationCS264: Beyond Worst-Case Analysis Lecture #18: Smoothed Complexity and Pseudopolynomial-Time Algorithms
CS264: Beyond Worst-Case Analysis Lecture #18: Smoothed Complexity and Pseudopolynomial-Time Algorithms Tim Roughgarden March 9, 2017 1 Preamble Our first lecture on smoothed analysis sought a better theoretical
More informationImproved dynamic programming and approximation results for the Knapsack Problem with Setups
Improved dynamic programming and approximation results for the napsack Problem with Setups Ulrich Pferschy Rosario Scatamacchia Abstract We consider the 0 1 napsack Problem with Setups (PS). Items are
More informationCS 374: Algorithms & Models of Computation, Spring 2017 Greedy Algorithms Lecture 19 April 4, 2017 Chandra Chekuri (UIUC) CS374 1 Spring / 1
CS 374: Algorithms & Models of Computation, Spring 2017 Greedy Algorithms Lecture 19 April 4, 2017 Chandra Chekuri (UIUC) CS374 1 Spring 2017 1 / 1 Part I Greedy Algorithms: Tools and Techniques Chandra
More informationTopics in Theoretical Computer Science April 08, Lecture 8
Topics in Theoretical Computer Science April 08, 204 Lecture 8 Lecturer: Ola Svensson Scribes: David Leydier and Samuel Grütter Introduction In this lecture we will introduce Linear Programming. It was
More informationDiscrete Optimization 2010 Lecture 2 Matroids & Shortest Paths
Matroids Shortest Paths Discrete Optimization 2010 Lecture 2 Matroids & Shortest Paths Marc Uetz University of Twente m.uetz@utwente.nl Lecture 2: sheet 1 / 25 Marc Uetz Discrete Optimization Matroids
More information2 Making the Exponential Mechanism Exactly Truthful Without
CIS 700 Differential Privacy in Game Theory and Mechanism Design January 31, 2014 Lecture 3 Lecturer: Aaron Roth Scribe: Aaron Roth Privacy as a Tool for Mechanism Design for arbitrary objective functions)
More informationCS264: Beyond Worst-Case Analysis Lecture #15: Smoothed Complexity and Pseudopolynomial-Time Algorithms
CS264: Beyond Worst-Case Analysis Lecture #15: Smoothed Complexity and Pseudopolynomial-Time Algorithms Tim Roughgarden November 5, 2014 1 Preamble Previous lectures on smoothed analysis sought a better
More informationGreedy vs Dynamic Programming Approach
Greedy vs Dynamic Programming Approach Outline Compare the methods Knapsack problem Greedy algorithms for 0/1 knapsack An approximation algorithm for 0/1 knapsack Optimal greedy algorithm for knapsack
More informationData Structures in Java
Data Structures in Java Lecture 21: Introduction to NP-Completeness 12/9/2015 Daniel Bauer Algorithms and Problem Solving Purpose of algorithms: find solutions to problems. Data Structures provide ways
More informationSolutions to Exercises
1/13 Solutions to Exercises The exercises referred to as WS 1.1(a), and so forth, are from the course book: Williamson and Shmoys, The Design of Approximation Algorithms, Cambridge University Press, 2011,
More informationAPTAS for Bin Packing
APTAS for Bin Packing Bin Packing has an asymptotic PTAS (APTAS) [de la Vega and Leuker, 1980] For every fixed ε > 0 algorithm outputs a solution of size (1+ε)OPT + 1 in time polynomial in n APTAS for
More information1 Submodular functions
CS 369P: Polyhedral techniques in combinatorial optimization Instructor: Jan Vondrák Lecture date: November 16, 2010 1 Submodular functions We have already encountered submodular functions. Let s recall
More informationCS2223 Algorithms D Term 2009 Exam 3 Solutions
CS2223 Algorithms D Term 2009 Exam 3 Solutions May 4, 2009 By Prof. Carolina Ruiz Dept. of Computer Science WPI PROBLEM 1: Asymptoptic Growth Rates (10 points) Let A and B be two algorithms with runtimes
More informationA PTAS for the Uncertain Capacity Knapsack Problem
Clemson University TigerPrints All Theses Theses 12-20 A PTAS for the Uncertain Capacity Knapsack Problem Matthew Dabney Clemson University, mdabney@clemson.edu Follow this and additional works at: https://tigerprints.clemson.edu/all_theses
More informationScheduling Parallel Jobs with Linear Speedup
Scheduling Parallel Jobs with Linear Speedup Alexander Grigoriev and Marc Uetz Maastricht University, Quantitative Economics, P.O.Box 616, 6200 MD Maastricht, The Netherlands. Email: {a.grigoriev, m.uetz}@ke.unimaas.nl
More information