Lecture Notes CS:5360 Randomized Algorithms Lecture 20 and 21: Nov 6th and 8th, 2018 Scribe: Qianhang Sun

Similar documents
Lecture 10: Lovász Local Lemma. Lovász Local Lemma

NP-Hardness reductions

Lecture 18: PCP Theorem and Hardness of Approximation I

Lecture 1 : Probabilistic Method

n Boolean variables: x 1, x 2,...,x n 2 {true,false} conjunctive normal form:

Lecture 24 : Even more reductions

CSE 3500 Algorithms and Complexity Fall 2016 Lecture 25: November 29, 2016

Essential facts about NP-completeness:

Advanced Algorithms 南京大学 尹一通

CS21 Decidability and Tractability

Limits to Approximability: When Algorithms Won't Help You. Note: Contents of today s lecture won t be on the exam

Lecture 10 Algorithmic version of the local lemma

Randomness and Computation March 13, Lecture 3

Approximation algorithm for Max Cut with unit weights

Lecture 11: Generalized Lovász Local Lemma. Lovász Local Lemma

College of Computer & Information Science Fall 2009 Northeastern University 22 September 2009

Notes for Lecture 2. Statement of the PCP Theorem and Constraint Satisfaction

Lecture 21 (Oct. 24): Max Cut SDP Gap and Max 2-SAT

Lecture 13, Fall 04/05

Algorithms and Theory of Computation. Lecture 22: NP-Completeness (2)

20.1 2SAT. CS125 Lecture 20 Fall 2016

Topics in Probabilistic Combinatorics and Algorithms Winter, Basic Derandomization Techniques

Lecture 18: More NP-Complete Problems

1 Primals and Duals: Zero Sum Games

CS Introduction to Complexity Theory. Lecture #11: Dec 8th, 2015

Intro to Theory of Computation

Lecture 23: More PSPACE-Complete, Randomized Complexity

Polynomial-Time Reductions

Lecture 10: Hardness of approximating clique, FGLSS graph

PCPs and Inapproximability Gap-producing and Gap-Preserving Reductions. My T. Thai

Lecture 13 March 7, 2017

CS151 Complexity Theory. Lecture 13 May 15, 2017

1 Agenda. 2 History. 3 Probabilistically Checkable Proofs (PCPs). Lecture Notes Definitions. PCPs. Approximation Algorithms.

princeton univ. F 17 cos 521: Advanced Algorithm Design Lecture 6: Provable Approximation via Linear Programming

CSCI 1590 Intro to Computational Complexity

NP-Completeness Part II

U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 8 Luca Trevisan September 19, 2017

More NP-Complete Problems

UC Berkeley CS 170: Efficient Algorithms and Intractable Problems Handout 22 Lecturer: David Wagner April 24, Notes 22 for CS 170

NP-Completeness Part II

P P P NP-Hard: L is NP-hard if for all L NP, L L. Thus, if we could solve L in polynomial. Cook's Theorem and Reductions

NP-complete Problems

CS21 Decidability and Tractability

NP and NP-Completeness

1 More finite deterministic automata

An Improved Upper Bound for SAT

CS6999 Probabilistic Methods in Integer Programming Randomized Rounding Andrew D. Smith April 2003

NP and Computational Intractability

CS Lecture 29 P, NP, and NP-Completeness. k ) for all k. Fall The class P. The class NP

U.C. Berkeley CS294: Beyond Worst-Case Analysis Handout 2 Luca Trevisan August 29, 2017

CS154, Lecture 18: 1

Lecture 24 Nov. 20, 2014

-bit integers are all in ThC. Th The following problems are complete for PSPACE NPSPACE ATIME QSAT, GEOGRAPHY, SUCCINCT REACH.

Approximations for MAX-SAT Problem

Provable Approximation via Linear Programming

Theory of Computation CS3102 Spring 2014 A tale of computers, math, problem solving, life, love and tragic death

Algorithm Design and Analysis

Approximations for MAX-SAT Problem

NP-Complete Problems. More reductions

Advanced Algorithms (XIII) Yijia Chen Fudan University

Lecture 12: Interactive Proofs

SAT, Coloring, Hamiltonian Cycle, TSP

Algorithms Design & Analysis. Approximation Algorithm

Chapter 2. Reductions and NP. 2.1 Reductions Continued The Satisfiability Problem (SAT) SAT 3SAT. CS 573: Algorithms, Fall 2013 August 29, 2013

Lecture 5: January 30

Lecture 24: April 12

CSCI3390-Second Test with Solutions

NP Completeness. CS 374: Algorithms & Models of Computation, Spring Lecture 23. November 19, 2015

Theorem (Special Case of Ramsey s Theorem) R(k, l) is finite. Furthermore, it satisfies,

Show that the following problems are NP-complete

Adleman Theorem and Sipser Gács Lautemann Theorem. CS422 Computation Theory CS422 Complexity Theory

COP 4531 Complexity & Analysis of Data Structures & Algorithms

Classes of Problems. CS 461, Lecture 23. NP-Hard. Today s Outline. We can characterize many problems into three classes:

SAT, NP, NP-Completeness

Lecture 20: PSPACE. November 15, 2016 CS 1010 Theory of Computation

CS154, Lecture 15: Cook-Levin Theorem SAT, 3SAT

8. INTRACTABILITY I. Lecture slides by Kevin Wayne Copyright 2005 Pearson-Addison Wesley. Last updated on 2/6/18 2:16 AM

COMP 382. Unit 10: NP-Completeness

How to Cope with NP-Completeness

Approximation Preserving Reductions

Robin Moser makes Lovász Local Lemma Algorithmic! Notes of Joel Spencer

Algorithm Design and Analysis

Design and Analysis of Algorithms

PRAMs. M 1 M 2 M p. globaler Speicher

The Probabilistic Method

CS151 Complexity Theory. Lecture 15 May 22, 2017

Computational Intractability 2010/4/15. Lecture 2

CS145: Probability & Computing Lecture 24: Algorithms

NP Complete Problems. COMP 215 Lecture 20

Geometric Steiner Trees

Global Optima from Local Algorithms

U.C. Berkeley CS278: Computational Complexity Professor Luca Trevisan August 30, Notes for Lecture 1

A An Overview of Complexity Theory for the Algorithm Designer

NP-Complete Reductions 1

Asymptotic Polynomial-Time Approximation (APTAS) and Randomized Approximation Algorithms

Lecture 5: Probabilistic tools and Applications II

Non-Approximability Results (2 nd part) 1/19

CSC 2429 Approaches to the P vs. NP Question and Related Complexity Questions Lecture 2: Switching Lemma, AC 0 Circuit Lower Bounds

Probabilistically Checkable Proofs. 1 Introduction to Probabilistically Checkable Proofs

PCP Theorem and Hardness of Approximation

Transcription:

1 Probabilistic Method Lecture Notes CS:5360 Randomized Algorithms Lecture 20 and 21: Nov 6th and 8th, 2018 Scribe: Qianhang Sun Turning the MaxCut proof into an algorithm. { Las Vegas Algorithm Algorithm Deterministic Algorithm 2 MaxCut Proof Theorem. crossing it. Derandomization { Pairwise Independence Method of Conditional Probabilities Let G = (V, E) be a graph with m edges. G has a cut with no less than m/2 edges Algorithm 1: las vegas algorithm of maxcut : 1 repeat 2 Throw each vertex independently v V into A or B with prob 1/2; 3 C edges crossing the (A, B) cut; 4 until C m/2; Analyze. Let p = P r( C m/2 ), then E[ C ] = = m c P r( C = c) c=0 m/2 1 ( m 2 c=0 c P r( C = c) + 1) (1 p) + m p c= m/2 c P r( C = c) Since E[ C ] = m 2, m 2 (m 2 1) (1 p) + m p m 2 1 p (m 2 1) + m p p 1 1 + m 2 Thus, the expected number of repetitions is O(m), which is, total running time is polynomial in repetition. 1

3 Derandomization Pairwise Independence. Let X v = { 1 if v falls in A 0 otherwise We assumed that {X v v V } are mutually independent. Suppose {X v v V } are only pairwise independent. i.e. P r(x u = 1 X v = 0) = P r(x u = 1) P r(x u = 1 X w = 1) P r(x u = 1) Is it still the case that E[ C ] = m 2? Yes. Constructing random variables that are pairwise independent. Let m 1, n = 2 m 1, suppose we are given m mutually independent (0 1) random variables Y 1, Y 2, Y 3 Y m. Let S 1, 2, m, S ø, X s = XORofY i s, i S. Since the number of such sets S is 2 m 1, we have (2 m 1) (0 1) random variables X s. Claim. {X s S subseteq1, 2, 3 m, S ø} are pairwise independent. P r(x s = 1) = 1 2 Figure 1: Pairwise Independent We need a random variable X v for each vertex v V. Thus we need V pairwise independent random variables log 2 V mutually independent random variables are needed. Since we only need log 2 V random bits, we can generate all possible settings of These in O( V ) time and get the entire space. Then we explore each cut in the sample space and pick a cut of size no less than m 2, which is guaranteed to exist. Method of Conditional Probabilities. Claim. There exists x k+1 {A, B}, E[C(A, B) x 1, x 2, x k+1 ] E[C(A, B) x 1, x 2, x k ] 2

Figure 2: Note that E[C(A, B) x 1, x 2, x k ] = size of a specific cut. {E[C(A, B) x 1, x 2, x k ], x i {A, B}} denotes the conditional expectation of C(A, B), the size of the cut, conditioned on v i falling into x i, for i = 1, 2, 3 k. Proof: E[C(A, B) x 1, x 2, x k ] = 1 2 E[C(A, B) x 1, x 2, x k, A] + 1 2 E[C(A, B) x 1, x 2, x k, B] Algorithm step at node E[C(A, B) x 1, x 2, x k ]: - Calculate E[C(A, B) x 1, x 2, x k, A] and E[C(A, B) x 1, x 2, x k, B] - Travel to the child node with larger expectation. How to calculate E[C(A, B) x 1, x 2, x k, A]? (1) Count number of edges with both end points fixed that cross the cut. (2) The answer = count in step(1) + 1 2 (remaining edges) Note that the remaining edges term is the same independent of which set v k+1 is assigned to. Thus, v k+1 needs to be placed in a set A or B that maximizes the number of edges crossing the cut. Theorem: The greedy algorithm for MaxCut produces a cut of size no less than m 2 4 Lovasz Local Lemma - Gem of the probabilistic method - 1975 Lovasz & Erdos: on hypergraph coloring We have a collection B 1, B 2, B n of bad events. Goal: Pr(no bad event occurs) > 0. i.e. P r( n B i=0 i ) > 0 There exists a good element in the sample space. How to show P r( n B i=0 i ) > 0? 3

Approache 1: P r(b i ) is very small, then n P r( i=0 n B i ) = 1 P r( B i ) 1 i=0 i=1 n P r(b i ) If n i=1 P r(b i) < 1,then P r( n B i=0 i ) > 0. So for example, if P r(b i ) < 1 n, then this holds. Approache 2: Independence n P r( i=0 n B i ) = P r( B i ) (by independence) i=1 If P r( B i ) > 0 for all i, we are done. In settings where P r( B i ) is not too small and B i s are not mutually independent, we use Lovasz Local Lemma. Idea: - P r(b i ) p - p is not too small, but small enough relative to the dependencies among the B i s. Definition: Let B 1, B 2, B n be events, A directed graph G = (V, E) with V = {1, 2, n} is a dependency graph of the events if every event B i is mutually independent of {B j (i, j) / E} Example: Consider an experiment in which we toss two fair, independent coins. E 1 : first coin toss is Head E 2 : second coin toss is Head E e : two coin tosses are identical Figure 3: Relation of E 1, E 2 and E 3 Is E 1 mutually independent with respect to {E 2, E 3 }? No. See edge 1. Also, dependency graph is not unique. 4

Lovasz Local Lemma: Let B 1, B 2, B n be events such that: (1) P r(b i ) p, for i=1, 2, n. (2) Maximum outdegree of a dependency graph of B 1, B 2, B n is d. (3) 4pd 1 (i.e. p 1 4 ) Then P r( n B i=0 i ) > 0). Example 1: We are given a n-vertex cycle. We want to properly color the vertices, i.e. no two adjacent vertices have the same color. Then, how many colors are suffice to choose? 3 colors. Using Lovasz Local Lemma we will show that 8 or 9 colors suffice. Let us start with a palette of c colors. Each vertex is colored uniformly random using a color from the palette. Then, good event = all pairs of adjacent vertices choose different colors. Let e 1, e 2 e n be the edges of the cycle. B ei = both endpoints of e i have the same color. Good event = n i=1 B ei and P r(b ei ) = 1 c. What about dependencies among B e i s? We need Figure 4: Relation of e 1, e 2 e n. In this figure, it is easy to tell d=2. Instead of d = 2, we had d = 1, then c 4. 4pd 1 4 1 c 2 1 c 8 Example 2: K-SAT An instance of SAT is a boolean formula in CNF. For example: Literal {}}{ ( X 1 }{{ X } 2 ) (X 2 X 3 X 4 ) ( }{{} X 1 X 4 ) }{{} Clause Clause Clause 5

k-sat = special case of SAT in which each clause has exactly k literals. Is the given instance of k-sat satisfiable? Theorem: satisfiable. If each variable appears at most T := 2k 4k clauses, then the given instance of k-sat is Proof: via Lovasz Local Lemma For each variable x i, set it independently to true or false with probability 1 2 each. For each clause C, define B c = event that C is False. P r(b c ) = 1 2 k Figure 5: k-sat Thus, d 2k 4, Then 4 1 2 k 2k 4 = 1 Lovasz Local Lemma 6