A Note on the Budgeted Maximization of Submodular Functions

Similar documents
Near-optimal Nonmyopic Value of Information in Graphical Models

9. Submodular function optimization

Efficient Information Planning in Graphical Models

Optimization of Submodular Functions Tutorial - lecture I

Maximizing Submodular Set Functions Subject to Multiple Linear Constraints

Network Design and Game Theory Spring 2008 Lecture 2

Approximation Algorithms for Maximum. Coverage and Max Cut with Given Sizes of. Parts? A. A. Ageev and M. I. Sviridenko

Maximizing a Submodular Function with Viability Constraints

CSE541 Class 22. Jeremy Buhler. November 22, Today: how to generalize some well-known approximation results

1 Submodular functions

Revisiting the Greedy Approach to Submodular Set Function Maximization

Maximization of Submodular Set Functions

Submodularity in Machine Learning

Multi-document Summarization via Budgeted Maximization of Submodular Functions

Streaming Algorithms for Submodular Function Maximization

On Subset Selection with General Cost Constraints

Applications of Submodular Functions in Speech and NLP

Constrained Maximization of Non-Monotone Submodular Functions

arxiv: v1 [math.oc] 3 Jan 2019

Active Learning and Optimized Information Gathering

a 1 a 2 a 3 a 4 v i c i c(a 1, a 3 ) = 3

Maximizing Non-monotone Submodular Set Functions Subject to Different Constraints: Combined Algorithms

1 Overview. 2 Multilinear Extension. AM 221: Advanced Optimization Spring 2016

Optimal Sensor Placement and Scheduling with Value of Information for Spatio-Temporal Infrastructure System Management

Submodular Functions Properties Algorithms Machine Learning

arxiv: v1 [math.oc] 1 Jun 2015

The Power of Local Search: Maximum Coverage over a Matroid

Performance Guarantees for Information Theoretic Active Inference

Submodular and Linear Maximization with Knapsack Constraints. Ariel Kulik

Learning symmetric non-monotone submodular functions

Structured Robust Submodular Maximization: Offline and Online Algorithms

Approximability of Adaptive Seeding under Knapsack Constraints

On the Complexity of Budgeted Maximum Path Coverage on Trees

K-center Hardness and Max-Coverage (Greedy)

The Power of Local Search: Maximum Coverage over a Matroid

An Online Algorithm for Maximizing Submodular Functions

Monotone Submodular Maximization over a Matroid

ESPN: Efficient Server Placement in Probabilistic Networks with Budget Constraint

Robust and MaxMin Optimization under Matroid and Knapsack Uncertainty Sets

APPROXIMATION ALGORITHMS RESOLUTION OF SELECTED PROBLEMS 1

Submodular Functions and Their Applications

Structure Learning: the good, the bad, the ugly

A First Course in Combinatorial Optimization. JON LEE IBM T.J. Watson Research Center, Yorktown Heights, New York

Submodularity and curvature: the optimal algorithm

Latent Tree Approximation in Linear Model

Symmetry and hardness of submodular maximization problems

Submodular Maximization with Cardinality Constraints

A (k + 3)/2-approximation algorithm for monotone submodular k-set packing and general k-exchange systems

Computing and Communications 2. Information Theory -Entropy

Maximization Problems with Submodular Objective Functions

This means that we can assume each list ) is

Adaptive Submodularity: Theory and Applications in Active Learning and Stochastic Optimization

Homework 1 Due: Thursday 2/5/2015. Instructions: Turn in your homework in class on Thursday 2/5/2015

Filtered Search for Submodular Maximization with Controllable Approximation Bounds

Some Fixed-Point Results for the Dynamic Assignment Problem

Minimization of Symmetric Submodular Functions under Hereditary Constraints

An Introduction to Submodular Functions and Optimization. Maurice Queyranne University of British Columbia, and IMA Visitor (Fall 2002)

On Lagrangian Relaxation and Subset Selection Problems

Submodular Secretary Problem and Extensions

On the knapsack closure of 0-1 Integer Linear Programs. Matteo Fischetti University of Padova, Italy

HEURISTIC METHODS FOR DESIGNING UNIMODULAR CODE SEQUENCES WITH PERFORMANCE GUARANTEES

Computing Solution Concepts of Normal-Form Games. Song Chong EE, KAIST

Discrete Optimization in Machine Learning. Colorado Reed

MATHEMATICAL ENGINEERING TECHNICAL REPORTS. On the Pipage Rounding Algorithm for Submodular Function Maximization A View from Discrete Convex Analysis

On Greedy Maximization of Entropy

A Heuristic Algorithm for General Multiple Nonlinear Knapsack Problem

Distributed Submodular Maximization

Chapter 1. Preliminaries

8 Knapsack Problem 8.1 (Knapsack)

Optimal Budget Allocation: Theoretical Guarantee and Efficient Algorithm

Stochastic Submodular Cover with Limited Adaptivity

EE595A Submodular functions, their optimization and applications Spring 2011

c 2014 Society for Industrial and Applied Mathematics

ON THE EQUIVALENCE OF CONGLOMERABILITY AND DISINTEGRABILITY FOR UNBOUNDED RANDOM VARIABLES

CS 322: (Social and Information) Network Analysis Jure Leskovec Stanford University

Submodular Maximization and Diversity in Structured Output Spaces

Polyhedral Results for A Class of Cardinality Constrained Submodular Minimization Problems

Information Theory + Polyhedral Combinatorics

A necessary and sufficient condition for the existence of a spanning tree with specified vertices having large degrees

21. Solve the LP given in Exercise 19 using the big-m method discussed in Exercise 20.

On the knapsack closure of 0-1 Integer Linear Programs

Optimizing Ratio of Monotone Set Functions

arxiv: v1 [cs.ds] 14 Dec 2017

Parsimonious Tomography: Optimizing Cost-Identifiability Trade-off for Probing-based Network Monitoring

More Approximation Algorithms

Introduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information.

Dept. of Linguistics, Indiana University Fall 2015

A Polytope for a Product of Real Linear Functions in 0/1 Variables

Approximation Algorithms for Online Weighted Rank Function Maximization under Matroid Constraints

n-step mingling inequalities: new facets for the mixed-integer knapsack set

A robust APTAS for the classical bin packing problem

An Approximation Algorithm for MAX-2-SAT with Cardinality Constraint

arxiv: v2 [cs.ds] 28 Aug 2014

Approximating Submodular Functions. Nick Harvey University of British Columbia

Introduction to optimization and operations research

Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits

Machine Learning

Entropy and Ergodic Theory Lecture 4: Conditional entropy and mutual information

Fast Multi-Stage Submodular Maximization: Extended version

Transcription:

A Note on the udgeted Maximization of Submodular Functions Andreas Krause June 2005 CMU-CALD-05-103 Carlos Guestrin School of Computer Science Carnegie Mellon University Pittsburgh, PA 15213 Abstract Many set functions F in combinatorial optimization satisfy the diminishing returns property F A {X}) F A) F A {X}) F A ) for A A. Such functions are called submodular. A result from Nemhauser et.al. states that the problem of selecting k-element subsets maximizing a nondecreasing submodular function can be approximated with a constant factor 1 1/e) performance guarantee. Khuller et.al. showed that for the special submodular function involved in the MAX-COVER problem, this approximation result generalizes to a budgeted setting under linear nonnegative cost-functions. They proved a 1 1/ e) approximation guarantee for a modified Greedy algorithm, and show how a 1 1/e) guarantee can be achieved using partial enumeration. In this note, we ext their results to general submodular functions. Motivated by the problem of maximizing entropy in discrete graphical models, where the submodular objective cannot be evaluated exactly, we generalize our result to account for absolute errors. Revised on Nov 15 2011. Thanks to Giorgos Papachristoudis for helpful comments.

Keywords: Submodular functions; Optimization; Constraints; Entropy maximization

1 Introduction Many set functions F in combinatorial optimization satisfy the diminishing returns property F A X) F A) F A X) F A ) for A A, i.e. adding an element to a smaller set helps more than adding it to a larger set. Such functions are called submodular. Here and in the following, we write A X instead of A {X} to simplify notation. The submodular function motivating our research is the joint entropy HA) for a set of random variables A. The entropy of a distribution P : {x 1,..., x d } [0, 1] is defined as HP ) = k P x k ) log P x k ), measuring the number of bits required to encode {x 1,..., x d } [1]. If A is a set of discrete random variables A = {X 1,..., X n }, then their entropy HA) is defined as the entropy of their joint distribution. The conditional entropy HA ) for two subsets A, V is defined as HA ) = P a, b) log P a b), a dom A b dom measuring the expected uncertainty about variables A after variables are observed. Using the chain-rule of entropies [1], HA ) = HA ) + H), we can compute HA X) HA) = HX A). The information never hurts principle [1], HX A) HX A ) for all A A, proves submodularity of the entropy. In the discrete setting, HX A) is also always non-negative, hence the entropy is nondecreasing. In practice, a commonly used algorithm for selecting a set of variables with maximum entropy is to greedily select the next variable to observe as the most uncertain variable given the ones observed thus far: X k := argmax HX {X 1,... X k 1 }), 1.1) X which is again motivated by the chain-rule. It is no surprise that this problem has been tackled with heuristic approaches, since even the unit cost has been shown to be NP-hard for multivariate Gaussian distributions [3], and a related formulation has been shown to be NP PP -hard even for discrete distributions that can be represented by polytree graphical models [4]. Fortunately, a result from Nemhauser et.al. [5] states that the problem of selecting k-element subsets maximizing a nondecreasing submodular function can be approximated with a constant factor 1 1/e) performance guarantee, using the greedy algorithm as mentioned above. Khuller et.al. [2] showed that for the special submodular function involved in the MAX-COVER problem, this approximation result generalizes to a budgeted setting under linear nonnegative cost-functions. They proved a 1 1/ e) approximation guarantee for a modified Greedy algorithm, and show how a 1 1/e) guarantee can be achieved using partial enumeration. In this note, we ext their results to general submodular functions. A different proof of the 1 1/e) approximation guarantee for the partial enumeration algorithm has previously been presented by Sviridenko [6]. Motivated by the problem of maximizing entropy in discrete graphical models, where the conditional entropies in 1.1) can in general not be evaluated both exactly and efficiently [4], we generalize our result to account for absolute errors. Our derivations in the following sections closely follows the analysis presented in [2]. 2 udgeted maximization of submodular functions Let V be a finite set, and F : V R be a set function with F ) = 0. F is called submodular if F A X) F A) F A X) F A ) for all A V and X V \A. F is called non-decreasing if F A X) F A) 0 for all A V and X V \ A. The quantities F A; X) := F A X) F A) are called marginal increases of F with respect to A and X. Furthermore define a cost function c : V R +, associating a positive cost cx) with each element X V. We ext c linearly to sets: For A V define ca) = X A cx). 1

For a budget > 0, the budgeted maximization problem is to maximize OP T = argmax F A) 2.1) A V:cA) Note that the exclusion of zero cost does not incur loss of generality because since the submodular functions are nondecreasing. We refer to ca) = A as the unit-cost case. 3 A constant factor approximation In analogy to the unit-cost case discussed in [5], we analyze the greedy algorithm, where the greedy rule adds to set A the element X such that ˆF X G i 1 ; X) = max. X W\G i 1 cx i ) Khuller et.al. [2] prove that the simple greedy algorithm with this greedy selection rule has unbounded approximation ratio. They suggest a small modification, considering the best single element solution as alternative to the output of the naive greedy heuristic, which, as they prove, guarantees a constant factor approximation for the budgeted MAX-k-COVER problem. Their algorithm is stated here as Algorithm 1, and we ext their analysis to the case of general submodular functions. Motivated by the entropy maximization problem where we cannot efficiently evaluate the marginal increases F A; X) exactly [4], we only assume that we can evaluate ˆF A; X) such that ˆF A; X) F A; X) ε for some ε > 0. 1 Input: d > 0, > 0, W V Output: Selection A W begin A 1 :=argmax{f {X}) : X W,cX) }; A 2 := ; W := W; while W do foreach X W do X := ˆF A 2 ; X); X := argmax{ X /cx) : X W }; if ca 2 ) + cx ) then A 2 := A 2 X ; W := W \ X ; return argmax F A) A {A 1,A 2} Algorithm 1: Approximation algorithm for budgeted case. Let us consider the computation of the set A 2 in Algorithm 1. Renumber V = {X 1,..., X n } and define G 0 = and G i = {X 1,..., X i } such that F G i ) F G i 1 ) + ε cx i ) max Y F G i 1 Y ) F G) ε. cy ) The sequence G j ) j corresponds to the sequence of assignments to A 2, and is motivated by the simple greedy rule, adding, for a prior selection G i 1, the element X i such that ˆF G i 1 ; X) X i = max. X W\G i 1 cx i ) Let l = max{i : cg i ) } be the index corresponding to the iteration, where A 2 is last augmented in Line 1, and let c min = min X cx). Hence A 2 = G l. We first prove the following Theorem: 2

Theorem 1 adapted from [2]). Algorithm 1 achieves an 1 1/ ) e F OP T ) 2ε c min approximation for 2.1), using O W 2 ) evaluations of ˆF. To prove Theorem 1, we need two lemmas. Let w = OP T. Lemma 2 generalized from [2]). For i = 1,..., l + 1, it holds that F G i ) F G i 1 ) cx F OP T ) F G i 1)) ε 1 + wcx ) Proof. Using monotonicity of F, we have F OP T ) F G i 1 ) F OP T G i 1 ) F G i 1 ) = F OP T \ G i 1 G i 1 ) F G i 1 ) Assume OP T \ G i 1 = {Y 1,..., Y m }, and let for j = 1,..., m Z j = F G i 1 {Y 1,..., Y j }) F G i 1 {Y 1,..., Y j 1 }). Then F OP T ) F G i 1 ) m j=1 Z j. Now notice that Z j ε cy j ) F G i 1 Y j ) F G i 1 ) ε F G F G i 1 ) + ε cy j ) cx i ) using submodularity in the first and the greedy rule in the second inequality. Since m j=1 cy j) it holds that F OP T ) F G i 1 ) = m j=1 Z j F G F G i 1 ) + ε cx i ) + mε F G F G i 1 ) + ε cx i ) + wε Lemma 3 adapted from [2]). For i = 1,..., l + 1 it holds that [ i F G i ) 1 1 cx ) ] k) F OP T ) ) cx i ) + w ε. ) Proof. Let i = 1 for sake of induction. We need to prove that F G 1 ) cx1) F OP T ) cx + w ε. This follows from Lemma 2 and since cx i ) + w 1 + wcx. Now let i > 1. We have F G i ) = F G i 1 ) + [F G i ) F G i 1 )] F G i 1 ) + cx [F OP T ) F G i 1)] ε = 1 cx ) F G i 1 ) + cx F OP T ) ε 1 cx ) [ i 1 1 1 cx k) i = 1 1 cx k) i = 1 1 + wcx ) 1 + wcx ) ) ) F OP T ) cx i ) + w ) ) F OP T ) ε 1 + wcx 1 cx ) ) ) k) F OP T ) cx i ) + w ε using Lemma 2 in the first and the induction hypothesis in the second inequality. ) ] ε ) ) cx i ) + w ε 1 cx ) + cx F OP T ) ε 1 + wcx ) 3

From now on let β = 2 c min c min + w. Proof of Theorem 1. As observed by [2], it holds that if a 1,..., a n R + such that a i = αa and α 0, 1], the function 1 n ai i=1 1 A )) achieves its minimum at a 1 = = a n = αa n. We have [ l+1 F G l+1 ) 1 1 cx ) ] k) F OP T ) βε [ i 1 1 cx ) ] k) F OP T ) βε cg l+1 ) [ 1 1 1 ) ] l+1 F OP T ) βε l + 1 1 1 ) F OP T ) βε e where the first inequality follows from Lemma 3 the second inequality follows from the fact that cg l+1 ) > since it violates the budget, and the third inequality follows from above observation with α = 1. Furthermore note, that the violating increase F G l+1 ) F G l ) is bounded by F X ) for X = argmax X W F X), i.e. the second candidate solution considered by the modified greedy algorithm. Hence F G l ) + F X ) F G l+1 ) 1 1/e)F OP T ) βε and at least one of the values F X ) or F G l ) must be greater than or equal to 1 2 1 1/e)F OP T ) βε). Using an argument analog to the one of [2], this analysis can be further tightened. Note that if there is a single feasible, i.e., cost less than ) element X of value F X) 1 2F OP T ), Algorithm 1 will consider it, and produce a solution of value at least 1 2 F OP T ) 2ε. Now assume for all X, F X) < 1 2F OP T ). Let us consider two cases. In case cg l ) < 2, then for all X / G l it must hold that cx) > 2. Therefore, OP T \ G l 1. Suppose OP T \ G l = {X }. Since, by assumption, F X ) < OP T 2, by submodularity it must hold that F OP T G l ) > F OP T ) 2. y monotonicity it follows that F G l ) > F OP T ) 2. For the last remaining case, namely cg l ) 2, Lemma 3 and the observation above for α = 1 2 imply that F G l ) 1 1 l 1 cx ) ) k) F OP T ) βε 3.1) 1 1 ) ) l F OP T ) βε 1 1 ) e F OP T ) βε 3.2) 2l 4 An improved approximation guarantee To achieve the same performance guarantee of 1 1/e) which can be achieved for the unit-cost in the case of general submodular functions [5], Khuller et.al. [2] propose a partial enumeration heuristic which enumerates all subsets of up to d elements for some constant d > 0, and complements these subsets using the modified greedy algorithm Algorithm 1. They prove that this algorithm guarantees a 1 1/e) approximation for the budgeted MAX-k-COVER problem. The algorithm is stated below for general nondecreasing submodular functions: 4

Input: d > 0, > 0, W V Output: Selection A W begin A 1 :=argmax{f A) : A W, A <d,ca) } A 2 := ; foreach G W, G = d, cg) do W := W \ G; while W do foreach X W do X := ˆF G; X); X := argmax{ X /cx) : X W }; if cg) + cx ) then G := G X ; W := W \ X ; if F G) > F A 2 ) then A 2 := G return argmax F A) A {A 1,A 2} Algorithm 2: Approximation algorithm for budgeted case. Theorem 4 adapted from [2]). Algorithm 2 achieves an approximation guarantee of ) 2 1 1/e)F OP T ) ε for 2.1) if sets at least up to cardinality d = 3 are enumerated, using O W d+2 ) evaluations of ˆF. Proof of Theorem 4. Assume that OP T > d, otherwise the algorithm finds the exact optimum. Reorder OP T = {Y 1,..., Y m } such that Y i+1 = argmax F {Y 1,..., Y i, Y }) F {Y 1,..., Y i }), Y OP T and let = {Y 1,..., Y d }. Consider the iteration where the algorithm considers. Define the function c min F A) = F A ) F ). F is a nondecreasing submodular set function with F ) = 0, hence we can apply the modified greedy algorithm to it. Let A = {V 1,..., V l } be the result of the algorithm, where V i are chosen in sequence, let V l+1 be the first element from OP T \ which could not be added due to budget constraints, and let G = A. Per definition, F G) = F A) + F ). Let = F A V l+1 ) F A). Using Lemma 3, we find that F A) + 1 1/e)F OP T \ ) βε. Furthermore observe, since the elements in OP T are ordered, that F {Y 1,..., Y i }) F {Y 1,..., Y i 1 }) for 1 i d. Hence F ) d. Now we get F G) = F ) + F A) F ) + 1 1/e)F OP T \ ) βε F ) + 1 1/e)F OP T \ ) F ) βε d 1 1/d)F ) + 1 1/e)F OP T \ ) βε ut by definition, F ) + F OP T \ ) = F OP T ), and hence for d 3 F G) 1 1/e)F OP T ) βε. 5

5 Conclusions We presented an efficient approximation algorithm for the budgeted maximization of nondecreasing submodular set functions. We proved bounds on the absolute error which are incurred if the marginal increases can only be computed with an absolute error. We believe that our results are useful for the wide class of combinatorial optimization problems concerned with maximizing submodular functions. References [1] Thomas M. Cover and Joy A. Thomas, Elements of information theory, Wiley Interscience, 1991. [2] Samir Khuller, Anna Moss, and Joseph Naor, The budgeted maximum coverage problem, Inf Proc Let 70 1999), no. 1, 39 45. [3] Chunwa Ko, Jon Lee, and Maurice Queyranne, An exact algorithm for maximum entropy sampling, OR 43 1995), no. 4, 684 691. [4] A. Krause and C. Guestrin, Optimal nonmyopic value of information in graphical models - efficient algorithms and theoretical limits, Proc. of IJCAI, 2005. [5] G. Nemhauser, L. Wolsey, and M. Fisher, An analysis of the approximations for maximizing submodular set functions, Mathematical Programming 14 1978), 265 294. [6] Maxim Sviridenko, A note on maximizing a submodular set function subject to knapsack constraint, Operations Research Letters 32 2004), 41 43. 6