Junta Approximations for Submodular, XOS and Self-Bounding Functions

Similar documents
Optimal Bounds on Approximation of Submodular and XOS Functions by Juntas

Learning and Testing Submodular Functions

Learning Coverage Functions and Private Release of Marginals

Representation, Approximation and Learning of Submodular Functions Using Low-rank Decision Trees

Learning Pseudo-Boolean k-dnf and Submodular Functions

Agnostic Learning of Disjunctions on Symmetric Distributions

Learning symmetric non-monotone submodular functions

Approximating Submodular Functions. Nick Harvey University of British Columbia

Learning and Fourier Analysis

Learning and Fourier Analysis

Learning Combinatorial Functions from Pairwise Comparisons

Quantum boolean functions

Optimization of Submodular Functions Tutorial - lecture I

Lecture 5: February 16, 2012

Lecture 4: LMN Learning (Part 2)

Chebyshev Polynomials and Approximation Theory in Theoretical Computer Science and Algorithm Design

Linear sketching for Functions over Boolean Hypercube

3 Finish learning monotone Boolean functions

The Analysis of Partially Symmetric Functions

Faster Algorithms for Privately Releasing Marginals

Learning DNF Expressions from Fourier Spectrum

Lecture 10: Learning DNF, AC 0, Juntas. 1 Learning DNF in Almost Polynomial Time

Learning Submodular Functions

1 Last time and today

Active Testing! Liu Yang! Joint work with Nina Balcan,! Eric Blais and Avrim Blum! Liu Yang 2011

Upper Bounds on Fourier Entropy

A Noisy-Influence Regularity Lemma for Boolean Functions Chris Jones

Submodular Functions and Their Applications

THE CHOW PARAMETERS PROBLEM

Introduction to Computational Learning Theory

Query and Computational Complexity of Combinatorial Auctions

On the power and the limits of evolvability. Vitaly Feldman Almaden Research Center

Answering Many Queries with Differential Privacy

Embedding Hard Learning Problems Into Gaussian Space

CSE 291: Fourier analysis Chapter 2: Social choice theory

On Noise-Tolerant Learning of Sparse Parities and Related Problems

The Limitations of Optimization from Samples

From query complexity to computational complexity (for optimization of submodular functions)

Polynomial regression under arbitrary product distributions

Settling the Query Complexity of Non-Adaptive Junta Testing

Improved Approximation of Linear Threshold Functions

Guest Lecture: Average Case depth hierarchy for,, -Boolean circuits. Boaz Barak

Testing by Implicit Learning

Learning Set Functions with Limited Complementarity

With P. Berman and S. Raskhodnikova (STOC 14+). Grigory Yaroslavtsev Warren Center for Network and Data Sciences

New Results for Random Walk Learning

Fourier analysis of boolean functions in quantum computation

Harmonic Analysis of Boolean Functions

Unconditional Lower Bounds for Learning Intersections of Halfspaces

Lecture 7: Passive Learning

Lecture 9: March 26, 2014

A ROBUST KHINTCHINE INEQUALITY, AND ALGORITHMS FOR COMPUTING OPTIMAL CONSTANTS IN FOURIER ANALYSIS AND HIGH-DIMENSIONAL GEOMETRY

Monotone Submodular Maximization over a Matroid

On the Sensitivity of Cyclically-Invariant Boolean Functions

Active Learning: Disagreement Coefficient

Nearly Tight Bounds for Testing Function Isomorphism

Fast Rates for Estimation Error and Oracle Inequalities for Model Selection

Faster Private Release of Marginals on Small Databases

The Lovász Local Lemma: constructive aspects, stronger variants and the hard core model

Learning convex bodies is hard

On the Sensitivity of Cyclically-Invariant Boolean Functions

3.1 Decision Trees Are Less Expressive Than DNFs

How Low Can Approximate Degree and Quantum Query Complexity be for Total Boolean Functions?

Symmetry and hardness of submodular maximization problems

Chapter 9: Basic of Hypercontractivity

On the Fourier spectrum of symmetric Boolean functions

Lecture 29: Computational Learning Theory

PROPERTY TESTING LOWER BOUNDS VIA COMMUNICATION COMPLEXITY

Distribution Free Learning with Local Queries

arxiv: v3 [cs.ds] 22 Aug 2012

Learning Functions of Halfspaces Using Prefix Covers

6.842 Randomness and Computation April 2, Lecture 14

8.1 Polynomial Threshold Functions

Optimal Cryptographic Hardness of Learning Monotone Functions

Regularity, Boosting, and Efficiently Simulating Every High-Entropy Distribution

Lecture notes on OPP algorithms [Preliminary Draft]

arxiv: v1 [cs.cc] 29 Feb 2012

TTIC An Introduction to the Theory of Machine Learning. Learning from noisy data, intro to SQ model

Lecture 1: 01/22/2014

On Learning Monotone DNF under Product Distributions

An upper bound on l q norms of noisy functions

CS369E: Communication Complexity (for Algorithm Designers) Lecture #8: Lower Bounds in Property Testing

On the Complexity of Computing an Equilibrium in Combinatorial Auctions

On the Structure of Boolean Functions with Small Spectral Norm

1 Submodular functions

Quantum algorithms (CO 781/CS 867/QIC 823, Winter 2013) Andrew Childs, University of Waterloo LECTURE 13: Query complexity and the polynomial method

FKN Theorem on the biased cube

5.1 Learning using Polynomial Threshold Functions

Is Submodularity Testable? C. Seshadhri & Jan Vondrák

1 Learning Linear Separators

APPROXIMATION RESISTANCE AND LINEAR THRESHOLD FUNCTIONS

Learning Juntas. Elchanan Mossel CS and Statistics, U.C. Berkeley Berkeley, CA. Rocco A. Servedio MIT Department of.

Constrained Maximization of Non-Monotone Submodular Functions

A Composition Theorem for Conical Juntas

Learning Unions of ω(1)-dimensional Rectangles

Testing Properties of Boolean Functions

Learning DNF from Random Walks

arxiv: v1 [cs.gt] 4 Apr 2017

Geometric Series and the Ratio and Root Test

Chebyshev Polynomials, Approximate Degree, and Their Applications

Transcription:

Junta Approximations for Submodular, XOS and Self-Bounding Functions Vitaly Feldman Jan Vondrák IBM Almaden Research Center Simons Institute, Berkeley, October 2013 Feldman-Vondrák Approximations by Juntas 1 / 29

Junta approximations How can we simplify a function f : {0, 1} n R? R In this talk: How well can we approximate f by a function g of few variables? Def.: g approximates f within ɛ in L p, if f g p = (E[ f (x) g(x) p ]) 1/p ɛ (in this talk, the uniform distribution) Feldman-Vondrák Approximations by Juntas 2 / 29

Friedgut s Theorem Definition (Average sensitivity) The average sensitivity, or total influence, of f : {0, 1} n {0, 1} is Infl(f ) = n Pr x {0,1} n[f (x) f (x e i)]. i=1 Theorem (Friedgut 98) For any function f : {0, 1} n {0, 1} of average sensitivity Infl(f ) and every ɛ > 0, there is a function g depending on 2 O(Infl(f )/ɛ) variables such that f g 1 ɛ. Feldman-Vondrák Approximations by Juntas 3 / 29

Junta approximations of real-valued functions We investigate classes of real-valued functions f : {0, 1} n [0, 1]: submodular functions: f (x y) + f (x y) f (x) + f (y) XOS functions: f (x) = max i j a ijx j subadditive functions: f (x y) f (x) + f (y) self-bounding functions: f (x) i (f (x) f (x e i)) + Feldman-Vondrák Approximations by Juntas 4 / 29

Junta approximations of real-valued functions We investigate classes of real-valued functions f : {0, 1} n [0, 1]: submodular functions: f (x y) + f (x y) f (x) + f (y) XOS functions: f (x) = max i j a ijx j subadditive functions: f (x y) f (x) + f (y) self-bounding functions: f (x) i (f (x) f (x e i)) + Why these classes? Nice mathematical properties Role in game theory as valuation functions on bundles of goods [Balcan-Harvey 11] Can we learn valuations from random examples? Feldman-Vondrák Approximations by Juntas 4 / 29

Submodular functions Submodularity = property of diminishing returns. Let the marginal value of element j be j f (S) = f (S {j}) f (S). (we identify f (S) = f (1 S )) S T Definition: f is submodular, if for S T j cannot add more value to T than S. j j f (S) j f (T ) Equivalently: f (A B) + f (A B) f (A) + f (B). Feldman-Vondrák Approximations by Juntas 5 / 29

Subadditive and Fractionally Subadditive functions Definition: f is subadditive, if f (A B) f (A) + f (B) for all A, B. T S 1, S 2,... Definition: f is fractionally subadditive, if f (T ) α i f (S i ) whenever 1 T α i 1 Si. Feldman-Vondrák Approximations by Juntas 6 / 29

Subadditive and Fractionally Subadditive functions Definition: f is subadditive, if f (A B) f (A) + f (B) for all A, B. T S 1, S 2,... Definition: f is fractionally subadditive, if f (T ) α i f (S i ) whenever 1 T α i 1 Si. Definition: f is an XOS function, if f is a maximum over linear functions: f (x) = max i j a ijx j (a ij 0) Fact (for monotone functions with f ( ) = 0) Submodular Fract. Subadditive = XOS Subadditive Functions Feldman-Vondrák Approximations by Juntas 6 / 29

Self-bounding functions Definition: A function f : D n R + is a-self-bounding, if n i=1 (f (x) min y i D f (x 1,..., x i 1, y i, x i+1,..., x n )) af (x). Theorem: [Boucheron, Lugosi, Massart, 2000] 1-Lipschitz 1-self-bounding functions under product distributions are concentrated around E[f ] with standard deviation O( E[f ]) and Pr[f (X) < E[f ] t] < e t 2 /2E[f ], Pr[f (X) > E[f ] + t] > e t 2 /(2E[f ]+t). Feldman-Vondrák Approximations by Juntas 7 / 29

Self-bounding functions Definition: A function f : D n R + is a-self-bounding, if n i=1 (f (x) min y i D f (x 1,..., x i 1, y i, x i+1,..., x n )) af (x). Theorem: [Boucheron, Lugosi, Massart, 2000] 1-Lipschitz 1-self-bounding functions under product distributions are concentrated around E[f ] with standard deviation O( E[f ]) and Pr[f (X) < E[f ] t] < e t 2 /2E[f ], Pr[f (X) > E[f ] + t] > e t 2 /(2E[f ]+t). Fact XOS 1-Self-bounding functions. Submodular 2-Self-bounding functions. Feldman-Vondrák Approximations by Juntas 7 / 29

Overview of our function classes a-self-bounding self-bounding submodular XOS monotone submodular subadditive Feldman-Vondrák Approximations by Juntas 8 / 29

Related work (learning of submodular functions) [Balcan-Harvey 11] initiated the study of learning of submodular functions gave a learning algorithm for product distributions, using concentration properties of Lipschitz submodular functions proved a negative result for general distributions (no efficient learning within factors better than n 1/3 ) Feldman-Vondrák Approximations by Juntas 9 / 29

Related work (learning of submodular functions) [Balcan-Harvey 11] initiated the study of learning of submodular functions gave a learning algorithm for product distributions, using concentration properties of Lipschitz submodular functions proved a negative result for general distributions (no efficient learning within factors better than n 1/3 ) [Gupta-Hardt-Roth-Ullman 11] learning of submodular fn. with applications in differential privacy decomposition into n O(1/ɛ2) ɛ-lipschitz functions. Feldman-Vondrák Approximations by Juntas 9 / 29

Related work (learning of submodular functions) [Balcan-Harvey 11] initiated the study of learning of submodular functions gave a learning algorithm for product distributions, using concentration properties of Lipschitz submodular functions proved a negative result for general distributions (no efficient learning within factors better than n 1/3 ) [Gupta-Hardt-Roth-Ullman 11] learning of submodular fn. with applications in differential privacy decomposition into n O(1/ɛ2) ɛ-lipschitz functions. [Cheraghchi-Klivans-Kothari-Lee 12] learning based on Fourier analysis of submodular functions submodular fns are ɛ-approximable by polynomials of degree 1/ɛ 2 learning for uniform distributions, using n O(1/ɛ2) random examples Feldman-Vondrák Approximations by Juntas 9 / 29

Related work (cont d) [Rashodnikova-Yaroslavtsev 13, Blais-Onak-Servedio-Yaroslavtsev 13] learning/testing of discrete submodular functions (k possible values), using k O(k log k/ɛ) poly(n) value queries ɛ-approximation by a junta of size (k log 1 ɛ )Õ(k) Feldman-Vondrák Approximations by Juntas 10 / 29

Related work (cont d) [Rashodnikova-Yaroslavtsev 13, Blais-Onak-Servedio-Yaroslavtsev 13] learning/testing of discrete submodular functions (k possible values), using k O(k log k/ɛ) poly(n) value queries ɛ-approximation by a junta of size (k log 1 ɛ )Õ(k) [Feldman-Kothari-V. 13] ɛ-approximation of submodular functions by decision trees of depth O(1/ɛ 2 ), and hence juntas of size 2 O(1/ɛ2 ) PAC-learning using 2 poly(1/ɛ) poly(n) random examples (vs. [CKKL 12] n poly(1/ɛ) examples but in the agnostic setting) Feldman-Vondrák Approximations by Juntas 10 / 29

Related work (cont d) [Rashodnikova-Yaroslavtsev 13, Blais-Onak-Servedio-Yaroslavtsev 13] learning/testing of discrete submodular functions (k possible values), using k O(k log k/ɛ) poly(n) value queries ɛ-approximation by a junta of size (k log 1 ɛ )Õ(k) [Feldman-Kothari-V. 13] ɛ-approximation of submodular functions by decision trees of depth O(1/ɛ 2 ), and hence juntas of size 2 O(1/ɛ2 ) PAC-learning using 2 poly(1/ɛ) poly(n) random examples (vs. [CKKL 12] n poly(1/ɛ) examples but in the agnostic setting) QUESTIONS: Why is this restricted to submodular functions? Is the junta of size 2 O(1/ɛ2) related to Friedgut s Theorem? What are the best juntas that we can achieve? Feldman-Vondrák Approximations by Juntas 10 / 29

Our results [to appear in FOCS 13] Result 1: XOS and self-bounding functions can be ɛ-approximated in L 2 by juntas of size 2 O(1/ɛ2 ) This follows from a "real-valued Friedgut s theorem" Feldman-Vondrák Approximations by Juntas 11 / 29

Our results [to appear in FOCS 13] Result 1: XOS and self-bounding functions can be ɛ-approximated in L 2 by juntas of size 2 O(1/ɛ2 ) This follows from a "real-valued Friedgut s theorem" Result 2: Submodular fns can be ɛ-approximated by O( 1 ɛ 2 log 1 ɛ )-juntas Proof avoids Fourier analysis, uses concentration properties + "boosting lemma" from [Goemans-V. 04] Feldman-Vondrák Approximations by Juntas 11 / 29

Our results [to appear in FOCS 13] Result 1: XOS and self-bounding functions can be ɛ-approximated in L 2 by juntas of size 2 O(1/ɛ2 ) This follows from a "real-valued Friedgut s theorem" Result 2: Submodular fns can be ɛ-approximated by O( 1 ɛ 2 log 1 ɛ )-juntas Proof avoids Fourier analysis, uses concentration properties + "boosting lemma" from [Goemans-V. 04] Applications to learning: Submodular, XOS and monotone self-bounding functions can be PAC-learned in time 2 poly(1/ɛ) poly(n) within L 2 -error ɛ Submodular functions can be "PMAC"-learned in time 2 poly(1/ɛ) poly(n) within multiplicative error 1 + ɛ Feldman-Vondrák Approximations by Juntas 11 / 29

Overview of our junta approximations 2 O(a2 /ɛ 2 ) a-self-bounding 2 O(1/ɛ2 ) self-bounding Õ(1/ɛ 2 ) submodular 2 O(1/ɛ2 ) XOS Õ(1/ɛ 2 ) monotone submodular X subadditive Feldman-Vondrák Approximations by Juntas 12 / 29

No junta approximation for subadditive functions Example: any function f : {0, 1} n { 1 2, 1} is subadditive. 1 1 2 1 2 1 1 2 1 1 1 2 A, B; f (A B) 1 f (A) + f (B) Therefore, we can encode any function whatsoever, e.g. a parity function, which cannot be approximated by a junta. Feldman-Vondrák Approximations by Juntas 13 / 29

Plan Remaining plan of the talk: 1 Friedgut s theorem for real-valued functions 2 junta approximations for XOS and self-bounding functions 3 Improved junta approximation for submodular functions 4 Conclusions Feldman-Vondrák Approximations by Juntas 14 / 29

Friedgut s Theorem Friedgut s Theorem: for boolean functions f : {0, 1} n {0, 1} average sensitivity Infl(f ) ɛ-approx. by a junta of size 2O(Infl(f )/ɛ) Infl(f ) = n i=1 Pr x {0,1} n[f (x) f (x e i)] = S [n] S ˆf 2 (S) What should it say for real-valued functions? Feldman-Vondrák Approximations by Juntas 15 / 29

Friedgut s Theorem Friedgut s Theorem: for boolean functions f : {0, 1} n {0, 1} average sensitivity Infl(f ) ɛ-approx. by a junta of size 2O(Infl(f )/ɛ) Infl(f ) = n i=1 Pr x {0,1} n[f (x) f (x e i)] = S [n] S ˆf 2 (S) What should it say for real-valued functions? Natural extension of average sensitivity: Infl 2 (f ) = S [n] S ˆf 2 (S) = n i=1 E x {0,1} n[(f (x) f (x e i)) 2 ] But Friedgut s Theorem for this notion of average sensitivity is FALSE! as observed by [O Donnell-Servedio 07] Feldman-Vondrák Approximations by Juntas 15 / 29

Counterexample to Friedgut s Theorem? Counterexample for f : {0, 1} n [ 1, 1]: (from [O Donnell-Servedio 07]) +1 0 n n 2 n 2 n+ n 2 n n i=1 x i 1 x, i; f (x) f (x e i ) 1 n Infl 2 (f ) = n i=1 E[(f (x) f (x e i) 2 )] = O(1) so there should be an ɛ-approximate junta of size 2 O(1/ɛ)? but we need Ω(n) variables to approximate within a constant ɛ Feldman-Vondrák Approximations by Juntas 16 / 29

How to fix Friedgut s Theorem? [O Donnell-Servedio 07] prove a variant of Friedgut s theorem for discretized functions f : {0, 1} n [ 1, 1] δz. We don t know how to discretize while preserving submodularity etc. Feldman-Vondrák Approximations by Juntas 17 / 29

How to fix Friedgut s Theorem? [O Donnell-Servedio 07] prove a variant of Friedgut s theorem for discretized functions f : {0, 1} n [ 1, 1] δz. We don t know how to discretize while preserving submodularity etc. +1 0 n n 2 n 2 n+ n 2 n xi 1 Note: If we define Infl κ (f ) = n i=1 E[ f (x) f (x e i) κ ], then Infl 1 (f ) = n Θ(1/ n) = Θ( n). Feldman-Vondrák Approximations by Juntas 17 / 29

Friedgut s Theorem for real-valued functions Theorem Let f : {0, 1} n R. Then there exists a polynomial g of degree O(Infl 2 (f )/ɛ 2 ) depending on 2 O(Infl2 (f )/ɛ 2) poly(infl 1 (f )/ɛ) variables such that f g 2 ɛ. Feldman-Vondrák Approximations by Juntas 18 / 29

Friedgut s Theorem for real-valued functions Theorem Let f : {0, 1} n R. Then there exists a polynomial g of degree O(Infl 2 (f )/ɛ 2 ) depending on 2 O(Infl2 (f )/ɛ 2) poly(infl 1 (f )/ɛ) variables such that f g 2 ɛ. Notes: we could replace Infl 1 (f ) by Infl κ (f ) for κ < 2, but not Infl 2 (f ) for boolean functions, Infl 1 (f ) = Infl 2 (f ), so it doesn t matter we will show that for submodular, XOS and self-bounding functions, Infl 1 (f ) and Infl 2 (f ) are small Feldman-Vondrák Approximations by Juntas 18 / 29

Proof of Real-valued Friedgut Follow Friedgut s proof: Fourier analysis, hypercontractive inequality... Let d = 2Infl 2 (f )/ɛ 2 α = (ɛ 2 (κ 1) d /Infl κ (f )) κ/(2 κ), κ = 4/3 J = {i [n] : Infl κ i (f ) α} J = {S J, S d} Goal: S / J ˆf 2 (S) ɛ 2. Then, g(x) = S J ˆf (S)χ S (x) is an ɛ-approximation to f. Feldman-Vondrák Approximations by Juntas 19 / 29

Proof cont d The bound on ˆf S / J 2 (S) has two parts: ˆf 1 S >d 2 (S) 1 d S ˆf 2 (S) = 1 d Infl2 (f ) ɛ 2 /2 by the definition of d Feldman-Vondrák Approximations by Juntas 20 / 29

Proof cont d The bound on ˆf S / J 2 (S) has two parts: ˆf 1 S >d 2 (S) 1 d S ˆf 2 (S) = 1 d Infl2 (f ) ɛ 2 /2 by the definition of d ˆf 2 S J, S d 2 (S) ˆf i / J S d,i S 2 (S) (κ 1) 1 d i / J T κ 1 ( if ) 2 2 this part requires the hypercontractive inequality: T κ 1 (f ) 2 f κ where T ρ is the noise operator ( T ρ f (S) = ρ S ˆf (S)). Feldman-Vondrák Approximations by Juntas 20 / 29

Proof cont d The bound on ˆf S / J 2 (S) has two parts: ˆf 1 S >d 2 (S) 1 d S ˆf 2 (S) = 1 d Infl2 (f ) ɛ 2 /2 by the definition of d ˆf 2 S J, S d 2 (S) ˆf i / J S d,i S 2 (S) (κ 1) 1 d i / J T κ 1 ( if ) 2 2 this part requires the hypercontractive inequality: T κ 1 (f ) 2 f κ where T ρ is the noise operator ( T ρ f (S) = ρ S ˆf (S)). The difference: we need a bound on i / J if 2 κ = i / J (Inflκ i (f )) 2/κ, which does not follow from Infl 2 (f ) for real-valued functions. Finishing the proof: i / J (Inflκ i (f )) 2/κ α 2/κ 1 Infl κ i (f ) (κ 1) d ɛ 2. Feldman-Vondrák Approximations by Juntas 20 / 29

Application to self-bounding functions Recall: f is self-bounding if n (f (x) min xi f (x)) f (x). i=1 Feldman-Vondrák Approximations by Juntas 21 / 29

Application to self-bounding functions Recall: f is self-bounding if n (f (x) min xi f (x)) f (x). By double counting, i=1 Infl 1 (f ) = n E[ f (x) f (x e i ) ] = 2 i=1 n i=1 E[f (x) min f (x)] 2E[f (x)]. xi For f : {0, 1} n [0, 1], we get Infl 2 (f ) Infl 1 (f ) = O(1). Feldman-Vondrák Approximations by Juntas 21 / 29

Application to self-bounding functions Recall: f is self-bounding if n (f (x) min xi f (x)) f (x). By double counting, i=1 Infl 1 (f ) = n E[ f (x) f (x e i ) ] = 2 i=1 n i=1 E[f (x) min f (x)] 2E[f (x)]. xi For f : {0, 1} n [0, 1], we get Infl 2 (f ) Infl 1 (f ) = O(1). Corollary (of real-valued Friedgut) For any self-bounding (or submodular or XOS) function f : {0, 1} n [0, 1] and ɛ > 0, there is a polynomial g of degree d = O(1/ɛ 2 ) over 2 O(d) variables such that f g 2 ɛ. Feldman-Vondrák Approximations by Juntas 21 / 29

Lower bound for XOS functions Friedgut s Theorem: it is known that 2 Ω(1/ɛ) variables are necessary. Example: tribes function lower bound for XOS functions as well. 1 1 1 f (x) = max x i, x i,..., x i A 1 A 2 A b. i A 1 i A 2 i A b b = 2 1/ɛ disjoint blocks A j of size A j = 1/ɛ any junta smaller than 2 1/ɛ 1 misses 2 1/ɛ 1 blocks and cannot approximate f within ɛ Feldman-Vondrák Approximations by Juntas 22 / 29

Better juntas for submodular functions Theorem For every submodular function f : {0, 1} n [0, 1] and ɛ > 0, there is a function g depending on O( 1 ɛ 2 log 1 ɛ ) variables, such that f g 2 ɛ. Feldman-Vondrák Approximations by Juntas 23 / 29

Better juntas for submodular functions Theorem For every submodular function f : {0, 1} n [0, 1] and ɛ > 0, there is a function g depending on O( 1 ɛ 2 log 1 ɛ ) variables, such that f g 2 ɛ. Notes: this is tight up to the log factor; consider f (x) = ɛ 2 1/ɛ 2 i=1 x i in this sense, submodular functions are close to linear functions, while XOS/self-bounding functions are "more complicated" Feldman-Vondrák Approximations by Juntas 23 / 29

About the proof Inductive step: submodular function f of n variables a function of O( 1 ɛ 2 log n ɛ ) variables, approximating f within 1 2 ɛ n n 1 = 1 ɛ 2 log n ɛ n 2 = 1 ɛ 2 log n 1 ɛ n t = O( 1 ɛ 2 log 1 ɛ ) the process stops when n t = O( 1 ɛ 2 log 1 ɛ ) errors form a geometric series, converging to ɛ Feldman-Vondrák Approximations by Juntas 24 / 29

How to reduce n to J = 1 ɛ 2 log n ɛ For simplicity: assume f : {0, 1} n [0, 1] monotone submodular. Feldman-Vondrák Approximations by Juntas 25 / 29

How to reduce n to J = 1 ɛ 2 log n ɛ For simplicity: assume f : {0, 1} n [0, 1] monotone submodular. Idea: build J by including variables x i1, x i2, x i3,... that contribute significantly to the current set: Hopefully: E x {0,1} J [ i f (x)] = E x {0,1} J [f (x e i ) f (x)] > α. 1 we cannot include too many variables, because the function is bounded by [0, 1] 2 f is α-lipschitz in the variables that are not selected, and hence we can use concentration to argue that they can be ignored Feldman-Vondrák Approximations by Juntas 25 / 29

How to reduce n to J = 1 ɛ 2 log n ɛ For simplicity: assume f : {0, 1} n [0, 1] monotone submodular. Idea: build J by including variables x i1, x i2, x i3,... that contribute significantly to the current set: Hopefully: E x {0,1} J [ i f (x)] = E x {0,1} J [f (x e i ) f (x)] > α. 1 we cannot include too many variables, because the function is bounded by [0, 1] 2 f is α-lipschitz in the variables that are not selected, and hence we can use concentration to argue that they can be ignored BUT: f is α-lipschitz in the remaining variables only "in expectation"; we need a statement for most points in {0, 1} J and for all j / J Feldman-Vondrák Approximations by Juntas 25 / 29

The Boosting Lemma Lemma (Goemans,V. 2004) Let F {0, 1} n be down-closed (x y F x F) and σ(p) = Pr [x F] = p F (1 p) n F. x µ p F F Then σ(p) = (1 p) φ(p) where φ(p) is a non-decreasing function of p. Feldman-Vondrák Approximations by Juntas 26 / 29

The Boosting Lemma Lemma (Goemans,V. 2004) Let F {0, 1} n be down-closed (x y F x F) and σ(p) = Pr [x F] = p F (1 p) n F. x µ p F F Then σ(p) = (1 p) φ(p) where φ(p) is a non-decreasing function of p. Example: µ 1/2 F µ 1/k F σ( 1 2 ) 1 2 k = σ( 1 k ) (1 1 k )k 1 e Feldman-Vondrák Approximations by Juntas 26 / 29

How we find the small junta Algorithm: (for f monotone submodular) Initialize J :=, α ɛ 2, δ 1/ log n ɛ. Let J(δ) = each element of J independently with prob. δ. As long as i / J such that Pr[ i f (1 J(δ) ) > α] > 1/e, include i in J. Feldman-Vondrák Approximations by Juntas 27 / 29

How we find the small junta Algorithm: (for f monotone submodular) Initialize J :=, α ɛ 2, δ 1/ log n ɛ. Let J(δ) = each element of J independently with prob. δ. As long as i / J such that Pr[ i f (1 J(δ) ) > α] > 1/e, include i in J. Using the boosting lemma: If we did not include i in the final set J, then Pr x J(δ) [ i f (x) > α] 1/e, and hence Pr x J(1/2) [ i f (x) > α] (1/2) 1/δ ɛ/n. Feldman-Vondrák Approximations by Juntas 27 / 29

How we find the small junta Algorithm: (for f monotone submodular) Initialize J :=, α ɛ 2, δ 1/ log n ɛ. Let J(δ) = each element of J independently with prob. δ. As long as i / J such that Pr[ i f (1 J(δ) ) > α] > 1/e, include i in J. Using the boosting lemma: If we did not include i in the final set J, then Pr x J(δ) [ i f (x) > α] 1/e, and hence Pr x J(1/2) [ i f (x) > α] (1/2) 1/δ ɛ/n. Union bound Pr x J(1/2) [ i / J; i f (x) > α] ɛ. Feldman-Vondrák Approximations by Juntas 27 / 29

Finishing the proof Accuracy of the junta: We found a set J such that with probability 1 ɛ over x {0, 1} J, the function g x (y) = f (x, y) is ɛ 2 -Lipschitz in y By concentration, g x is ɛ-approximated by its expectation. Hence, f is 2ɛ-approximated by its averaging-projection on {0, 1} J. Feldman-Vondrák Approximations by Juntas 28 / 29

Finishing the proof Accuracy of the junta: We found a set J such that with probability 1 ɛ over x {0, 1} J, the function g x (y) = f (x, y) is ɛ 2 -Lipschitz in y By concentration, g x is ɛ-approximated by its expectation. Hence, f is 2ɛ-approximated by its averaging-projection on {0, 1} J. Size of the junta: every time we include i J, we have E x J(δ) [ i f (x)] > α/e so we increase E[f (J(δ))] by αδ/e this can repeat at most O( 1 αδ ) = O( 1 log n ɛ 2 ɛ ) times. Feldman-Vondrák Approximations by Juntas 28 / 29

Concluding comments and questions Self-bounding functions are ɛ-approximated by 2 O(1/ɛ2) -juntas Submodular functions are ɛ-approximated by Õ(1/ɛ2 )-juntas We also have a (1 + ɛ)-multiplicative approximation except for ɛ-fraction of {0, 1} n, for monotone submodular functions, by a junta of size Õ(1/ɛ2 ). We don t know if such a junta exists for non-monotone submodular functions More on our learning algorithms and results: Vitaly Feldman on Oct 30. Feldman-Vondrák Approximations by Juntas 29 / 29