Approximating MAX-E3LIN is NP-Hard

Similar documents
Lecture 18: Inapproximability of MAX-3-SAT

Essential facts about NP-completeness:

+ ɛ Hardness for Max-E3Lin

Lec. 11: Håstad s 3-bit PCP

Limits to Approximability: When Algorithms Won't Help You. Note: Contents of today s lecture won t be on the exam

Lecture 10: Hardness of approximating clique, FGLSS graph

Lecture 22: Counting

1 AC 0 and Håstad Switching Lemma

2 Completing the Hardness of approximation of Set Cover

Approximation Algorithms and Hardness of Approximation May 14, Lecture 22

Lecture 17. In this lecture, we will continue our discussion on randomization.

Sublinear Algorithms Lecture 3

Non-Approximability Results (2 nd part) 1/19

CS151 Complexity Theory. Lecture 15 May 22, 2017

PCPs and Inapproximability Gap-producing and Gap-Preserving Reductions. My T. Thai

Lecture 5: February 16, 2012

Lecture 15: A Brief Look at PCP

Lecture 12 Scribe Notes

CSC 5170: Theory of Computational Complexity Lecture 13 The Chinese University of Hong Kong 19 April 2010

Approximation algorithm for Max Cut with unit weights

CPSC 536N: Randomized Algorithms Term 2. Lecture 9

CS 151 Complexity Theory Spring Solution Set 5

Expander Graphs and Their Applications (X) Yijia Chen Shanghai Jiaotong University

PGSS Discrete Math Solutions to Problem Set #4. Note: signifies the end of a problem, and signifies the end of a proof.

NP-problems continued

DD2446 Complexity Theory: Problem Set 4

PAC Learning. prof. dr Arno Siebes. Algorithmic Data Analysis Group Department of Information and Computing Sciences Universiteit Utrecht

MATH 22 HAMILTONIAN GRAPHS. Lecture V: 11/18/2003

Advanced Algorithms (XIII) Yijia Chen Fudan University

Approximation Resistance from Pairwise Independent Subgroups

CS264: Beyond Worst-Case Analysis Lecture #11: LP Decoding

/633 Introduction to Algorithms Lecturer: Michael Dinitz Topic: Matroids and Greedy Algorithms Date: 10/31/16

Factoring. there exists some 1 i < j l such that x i x j (mod p). (1) p gcd(x i x j, n).

Lecture 24: Approximate Counting

Reading 11 : Relations and Functions

Cosets and Lagrange s theorem

CMPSCI 250: Introduction to Computation. Lecture #22: From λ-nfa s to NFA s to DFA s David Mix Barrington 22 April 2013

Lecture 14 - P v.s. NP 1

Lecture 18: PCP Theorem and Hardness of Approximation I

PCP Soundness amplification Meeting 2 ITCS, Tsinghua Univesity, Spring April 2009

Two Query PCP with Sub-Constant Error

Today. Few Comments. PCP Theorem, Simple proof due to Irit Dinur [ECCC, TR05-046]! Based on some ideas promoted in [Dinur- Reingold 04].

Compute the Fourier transform on the first register to get x {0,1} n x 0.

Lecture 26: Arthur-Merlin Games

Lecture 8 (Notes) 1. The book Computational Complexity: A Modern Approach by Sanjeev Arora and Boaz Barak;

Hardness of Approximation

1 Primals and Duals: Zero Sum Games

6.045: Automata, Computability, and Complexity (GITCS) Class 17 Nancy Lynch

1 Computational problems

Quantum Computing: Foundations to Frontier Fall Lecture 3

1 Maintaining a Dictionary

CS1800: Mathematical Induction. Professor Kevin Gold

CS151 Complexity Theory. Lecture 14 May 17, 2017

Languages, regular languages, finite automata

Lecture 5. 1 Review (Pairwise Independence and Derandomization)

Umans Complexity Theory Lectures

Lecture Hardness of Set Cover

Lecture 20: conp and Friends, Oracles in Complexity Theory

Lecture 59 : Instance Compression and Succinct PCP s for NP

Graph coloring, perfect graphs

Notes for Lecture 2. Statement of the PCP Theorem and Constraint Satisfaction

Lecture 17 November 8, 2012

1 Reductions and Expressiveness

Where do pseudo-random generators come from?

MA554 Assessment 1 Cosets and Lagrange s theorem

Lecture 15 - NP Completeness 1

On the efficient approximability of constraint satisfaction problems

Lecture 29: Computational Learning Theory

Proving languages to be nonregular

Lecture 23: Alternation vs. Counting

Theory of Computer Science to Msc Students, Spring Lecture 2

NP-Completeness I. Lecture Overview Introduction: Reduction and Expressiveness

Monotone Submodular Maximization over a Matroid

Today: Linear Programming (con t.)

Section 20: Arrow Diagrams on the Integers

Lecture 22. m n c (k) i,j x i x j = c (k) k=1

Lecture 6. Today we shall use graph entropy to improve the obvious lower bound on good hash functions.

PCP Theorem and Hardness of Approximation

Lecture 23 : Nondeterministic Finite Automata DRAFT Connection between Regular Expressions and Finite Automata

Lecture 4: LMN Learning (Part 2)

A function is a special kind of relation. More precisely... A function f from A to B is a relation on A B such that. f (x) = y

Algorithms Reading Group Notes: Provable Bounds for Learning Deep Representations

Lecture 7: ɛ-biased and almost k-wise independent spaces

U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 8 Luca Trevisan September 19, 2017

Notes on Computer Theory Last updated: November, Circuits

Lecture 3: Reductions and Completeness

Week 3: Reductions and Completeness

princeton univ. F 17 cos 521: Advanced Algorithm Design Lecture 6: Provable Approximation via Linear Programming

CS 124 Math Review Section January 29, 2018

Lecture 5: Derandomization (Part II)

CSC 5170: Theory of Computational Complexity Lecture 9 The Chinese University of Hong Kong 15 March 2010

CS 6820 Fall 2014 Lectures, October 3-20, 2014

Lecture 16: Constraint Satisfaction Problems

Approximation Algorithms and Hardness of Approximation. IPM, Jan Mohammad R. Salavatipour Department of Computing Science University of Alberta

CS151 Complexity Theory. Lecture 6 April 19, 2017

CSE 190, Great ideas in algorithms: Pairwise independent hash functions

Discrete Mathematics and Probability Theory Fall 2014 Anant Sahai Note 7

Lecture 8: Linearity and Assignment Testing

6.842 Randomness and Computation April 2, Lecture 14

15-451/651: Design & Analysis of Algorithms October 31, 2013 Lecture #20 last changed: October 31, 2013

Transcription:

Approximating MAX-E3LIN is NP-Hard Evan Chen May 4, 2016 This lecture focuses on the MAX-E3LIN problem. We prove that approximating it is NP-hard by a reduction from LABEL-COVER. 1 Introducing MAX-E3LIN In the MAX-E3LIN problem, our input is a series of linear equations (mod 2) in n binary variables, each with three terms. Equivalently, one can think of this as ±1 variables and ternary products. The objective is to maximize the fraction of satisfied equations. Example 1 (Example of MAX-E3LIN instance) x 1 + x 3 + x 4 1 (mod 2) x 1 + x 2 + x 4 0 (mod 2) x 1 + x 2 + x 5 1 (mod 2) x 1 + x 3 + x 5 1 (mod 2) A diligent reader can check that we may obtain 3 4 x 1 x 3 x 4 = 1 x 1 x 2 x 4 = +1 x 1 x 2 x 5 = 1 x 1 x 3 x 5 = 1 but not 1. Remark 2. We immediately notice that If there s a solution with value 1, we can find it easily with F 2 linear algebra. It is always possible to get at least 1 2 by selecting all-zero or all-one. The theorem we will prove today is that these obvious observations are essentially the best ones possible! Our main result is that improving the above constants to 51% and 99%, say, is NP-hard. Theorem 3 (Hardness of MAX-E3LIN) The 1 2 + ε vs. 1 δ decision problem for MAX-E3LIN is NP-hard. This means it is NP-hard to decide whether an MAX-E3LIN instance has value 1 2 + ε or 1 δ (given it is one or the other). A direct corollary of this is approximating MAX-SATis also NP-hard. Corollary 4 The 7 8 + ε vs. 1 δ decision problem for MAX-SAT is NP-hard. 1

Evan Chen 3 Setup Remark 5. The constant 7 8 is optimal in light of a random assignment. In fact, one can replace 1 δ with δ, but we don t do so here. Proof. Given an equation a + b + c = 1 in MAX-E3LIN, we consider four formulas a b c, a b c, a b c, a b c. Either three or four of them are satisfied, with four occurring exactly when a + b + c = 0. One does a similar construction for a + b + c = 1. The hardness of MAX-E3LIN is relevant to the PCP theorem: using MAX-E3LIN gadgets, Håstad was able to prove a very strong version of the PCP theorem, in which the verifier merely reads just three bits of a proof! Theorem 6 (Håstad PCP) Let ε, δ > 0. We have NP PCP 1 +ε,1 δ(3, O(log n)). 2 In other words, any L NP has a (non-adaptive) verifier with the following properties. The verifier uses O(log n) random bits, and queries just three (!) bits. The acceptance condition is either a + b + c = 1 or a + b + c = 0. If x L, then there is a proof Π which is accepted with probability at least 1 δ. If x / L, then every proof is accepted with probability at most 1 2 + ε. 2 Label Cover We will prove our main result by reducing from the LABEL-COVER. Recall LABEL-COVER is played as follows: we have a bipartite graph G = U V, a set of keys K for vertices of U and a set of labels L for V. For every edge e = {u, v} there is a function π e : L K specifying a key k = π e (l) K for every label l L. The goal is to label the graph G while maximizing the number of edges e with compatible key-label pairs. Approximating LABEL-COVER is NP-hard: Theorem 7 (Hardness of LABEL-COVER) The η vs. 1 decision problem for LABEL-COVER is NP-hard for every η > 0, given K and L are sufficiently large in η. So for any η > 0, it is NP-hard to decide whether one can satisfy all edges or fewer than η of them. 3 Setup We are going to make a reduction of the following shape: LABEL-COVER MAX-E3LIN Soundness < η < 1 2 + ε Completeness = 1 1 δ 2

Evan Chen 3 Setup In words this means that Completeness : If the LABEL-COVER instance is completely satisfiable, then we get a solution of value 1 δ in the resulting MAX-E3LIN. Soundness : If the LABEL-COVER instance has value η, then we get a solution of value 1 2 + ε in the resulting MAX-E3LIN. Thus given an oracle for MAX-E3LIN decision, we can obtain η vs. 1 decision for LABEL- COVER, which we know is hard. The setup for this is quite involved, using a huge number of variables. Just to agree on some conventions: Definition 8 ( Long Code ). A K-indexed binary string x = (x k ) k is a ±1 sequence indexed by K. We can think of it as an element of {±1} K. An L-binary string y = (y l ) l is defined similarly. Now we initialize U 2 K + V 2 L variables: At every vertex u U, we will create 2 K binary variables, one for every K-indexed binary string. It is better to collect these variables into a function f u : {±1} K {±1}. Similarly, at every vertex v V, we will create 2 L binary variables, one for every L-indexed binary string, and collect these into a function g v : {±1} L {±1}. Picture: π = π e : L K v g v : {±1} L ±1 u f u : {±1} K ±1 Next we generate the equations. Here s the motivation: we want to do this in such a way that given a satisfying labelling for LABEL-COVER, nearly all the MAX-E3LIN equations can be satisfied. One idea is as follows: for every edge e, letting π = π e, Take a K-indexed binary string x = (x k ) k at random. Take an L-indexed binary string y = (y l ) l at random. Define the L-indexed binary z = (z l ) l string by z = ( x π(l) y l ). Write down the equation f u (x)g v (y)g v (z) = +1 for the MAX-E3LIN instance. Thus, assuming we had a valid coloring of the graph, we could let f u and g v be the dictator functions for the colorings. In that case, f u (x) = x π(l), g v (y) = y l, and g v (z) = x π(l) y l, so the product is always +1. Unfortunately, this has two fatal flaws: 3

Evan Chen 5 A Fourier Computation 1. This means a 1 instance of LABEL-COVER gives a 1 instance of MAX-E3LIN, but we need 1 δ to have a hope of working. 2. Right now we could also just set all variables to be +1. We fix this as follows, by using the following equations. Definition 9 (Equations of reduction). For every edge e, with π = π e, we alter the construction and say Let x = (x k ) k be and y = (y l ) l be random as before. Let n = (n l ) l be a random L-indexed binary string, drawn from a δ-biased distribution ( 1 with probability δ). And now define z = (z l ) l by z l = x π(l) y l n l. The n l represents noise bits, which resolve the first problem by corrupting a bit of z with probability δ. Write down one of the following two equations with 1 2 probability each: This resolves the second issue. This gives a set of O( E ) equations. f u (x)g v (y)g v (z) = +1 f u (x)g v (y)g v ( z) = 1. I claim this reduction works. So we need to prove the completeness and soundness claims above. 4 Proof of Completeness Given a labeling of G with value 1, as described we simply let f u and g v be dictator functions corresponding to this valid labelling. Then as we ve seen, we will pass 1 δ of the equations. 5 A Fourier Computation Before proving soundness, we will first need to explicitly compute the probability an equation above is satisfied. Remember we generated an equation for e based on random strings x, y, λ. For T L, we define Thus T maps subsets of L to subsets of K. πe odd (T ) = { k K π e 1 (k) T is odd }. Remark 10. Note that π odd (T ) T and that π odd (T ) if T is odd. 4

Evan Chen 6 Proof of Soundness Lemma 11 (Edge Probability) The probability that an equation generated for e = {u, v} is true is 1 2 + 1 2 T L T odd (1 2δ) T ĝ v (T ) 2 fu (π odd e (T )). Proof. Omitted for now... 6 Proof of Soundness We will go in the reverse direction and show (constructively) that if there is MAX-E3LIN instance has a solution with value 1 2 + 2ε, then we can reconstruct a solution to LABEL-COVER with value η. (The use of 2ε here will be clear in a moment). This process is called decoding. The idea is as follows: if S is a small set such that f u (S) is large, then we can pick a key from S at random for f u ; compare this with the dictator functions where f u (S) = 1 and S = 1. We want to do something similar with T. Here are the concrete details. Let Λ = log(1/ε) 2δ and η = ε3 be constants (the actual Λ 2 values arise later). Definition 12. We say that a nonempty set S K of keys is heavy for u if S Λ and fu (S) ε 2. Note that there are at most ε 2 heavy sets by Parseval. Definition 13. We say that a nonempty set T L of labels is e-excellent for v if T Λ and S = πe odd (T ) is heavy. In particular S so at least one compatible key-label pair is in S T. Notice that, unlike the case with S, the criteria for good in T actually depends on the edge e in question! This makes it easier to keys than to select labels. In order to pick labels, we will have to choose from a ĝ 2 v distribution. Lemma 14 (At least ε of T are excellent) For any edge e = {u, v}, at least ε of the possible T according to the distribution ĝ 2 v are e-excellent. Proof. Applying an averaging argument to the inequality (1 2δ) T ĝ v (T ) 2 fu (π odd (T )) 2ε T L T odd shows there is at least ε chance that T is odd and satisfies (1 2δ) T fu (S) ε where S = π odd e (T ). In particular, (1 2δ) T ε T Λ. Finally by Remark 10, we see S is heavy. 5

Evan Chen 6 Proof of Soundness Now, use the following algorithm. For every vertex u U, take the union of all heavy sets, say H = S. S heavy Pick a random key from H. Note that H Λε 2, since there are at most ε 2 heavy sets (by Parseval) and each has at most Λ elements. For every vertex v V, select a random set T according to the distribution ĝ v (T ) 2, and select a random element from T. I claim that this works. Fix an edge e. There is at least an ε chance that T is e-excellent. If it is, then there is at least one compatible pair in H T. Hence we conclude probability of success is at least 1 ε Λε 2 1 Λ = ε3 Λ 2 = η. (Addendum: it s pointed out to me this isn t quite right; the overall probability of the equation given by an edge e is 1 2 + ε, but this doesn t imply it for every edge. Thus one likely needs to do another averaging argument.) 6