A Switching Lemma Tutorial. Benjamin Rossman University of Toronto and NII, Tokyo

Size: px
Start display at page:

Download "A Switching Lemma Tutorial. Benjamin Rossman University of Toronto and NII, Tokyo"

Transcription

1 A Switching Lemma Tutorial Benjamin Rossman University of Toronto and NII, Tokyo

2 homage to Paul Beame s excellent A Switching Lemma Primer A Switching Lemma Tutorial* Benjamin Rossman University of Toronto and NII, Tokyo

3 Overview 1. Basic Concepts models of computa?on (AC 0, DNFs/CNFs, decision trees) random restric?on R p 2. The Classic Switching Lemma (Hastad 86) 3. PARITY Lower Bound 4. Affine RestricGons Tsei?n expander switching lemma (joint with Pitassi, Servedio and Tan)

4 Models of Computa?on AC 0 circuits / formulas DNFs / CNFs (i.e. depth-2 formulas) Decision trees Canonical decision tree of a DNF / CNF

5 AC 0 Circuits size = # of gates depth = # layers of gates x 1 x 1 x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5

6 AC 0 Circuits size = # of gates depth = # layers of gates The complexity class AC 0 refers to (sequences of) Boolean func?ons {0,1} n {0,1} computable by poly(n)-size, constant-depth circuits. x 1 x 1 x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5

7 Depth-2 (DNFs and CNFs) DNF = disjunc?ve normal form (OR-AND formula) CNF = conjunc?ve normal form (AND-OR formula) width = max # of variables in a term/clause k

8 Depth-2 (DNFs and CNFs) k-dnf = width-k DNF k-cnf = width-k CNF k

9 Decision Trees x 3 depth 3 x 1 x 7 x 4 x 7 x 1 x

10 Decision Trees The decision-tree depth of a Boolean func?on f : {0,1} n {0,1} denoted DT depth (f), is the minimum depth of a decision tree that computes f DT depth (PARITY n ) = DT depth (AND n ) = n DT depth (f) = 0 f is constant

11 Decision Tree to DNF ( x 3 x 1 x 4 ) ( x 3 x 1 x 7 ) (x 3 x 7 x 1 ) (x 3 x 7 x 8 ) x 1 x 3 x 7 x 4 x 7 x 1 x

12 Decision Tree to DNF ( x 3 x 1 x 4 ) x 3 x 1 x 7 x 4 x 7 x 1 x

13 Decision Tree to DNF ( x 3 x 1 x 4 ) ( x 3 x 1 x 7 ) x 3 x 1 x 7 x 4 x 7 x 1 x

14 Decision Tree to DNF ( x 3 x 1 x 4 ) ( x 3 x 1 x 7 ) (x 3 x 7 x 1 ) x 3 x 1 x 7 x 4 x 7 x 1 x

15 Decision Tree to DNF ( x 3 x 1 x 4 ) ( x 3 x 1 x 7 ) (x 3 x 7 x 1 ) (x 3 x 7 x 8 ) x 1 x 3 x 7 x 4 x 7 x 1 x

16 (x 1 x 2 x 3 ) ( x 2 x 4 ) (x 1 x 3 x 7 ) Canonical DT of a DNF

17 (x 1 x 2 x 3 ) ( x 2 x 4 ) (x 1 x 3 x 7 ) Canonical DT of a DNF

18 Canonical DT of a DNF (x 1 x 2 x 3 ) x 1 x 2 x 3 1

19 Canonical DT of a DNF (x 1 x 2 x 3 ) ( x 2 x 4 ) x 1 x 2 x 4 1 x 4 1 x 2 x 2 x 4 1 x 3 x 2 1

20 Canonical DT of a DNF (x 1 x 2 x 3 ) ( x 2 x 4 ) x 1 x 2 x 4 1 x 4 1 x 2 x 2 x 4 1 x 3 x 2 1

21 Canonical DT of a DNF (x 1 x 2 x 3 ) ( x 2 x 4 ) x 1 x 2 x 4 1 x 4 x 2 1 x 3 1

22 Canonical DT of a DNF (x 1 x 2 x 3 ) ( x 2 x 4 ) x 1 (x 1 x 3 x 7 ) x 2 x x 3 x 7 x 4 x 2 1 x 7 1 x 3 1

23 Canonical DT of a DNF (x 1 x 2 x 3 ) ( x 2 x 4 ) x 1 (x 1 x 3 x 7 ) x 2 x x 3 x 7 0 x 4 0 x 2 1 x 7 1 x 3 0 1

24 Restric?ons a.k.a. func?ons [n] {0,1, } a.k.a. subcubes of {0,1} n

25 Restric?ons Consider a Boolean func?on f : {0,1} n {0,1} A restric3on (w.r.t. the variables of f) is a func?on R : [n] {0,1, }

26 Restric?ons Consider a Boolean func?on f : {0,1} n {0,1} A restric3on (w.r.t. the variables of f) is a func?on R : [n] {0,1, } [n] = {1,,n}, which we iden?fy with the variables x 1,,x n of f(x)

27 Restric?ons Consider a Boolean func?on f : {0,1} n {0,1} A restric3on (w.r.t. the variables of f) is a func?on R : [n] {0,1, } R(i) = 0 or 1 means that the variable x i is fixed to 0 or 1

28 Restric?ons Consider a Boolean func?on f : {0,1} n {0,1} A restric3on (w.r.t. the variables of f) is a func?on R : [n] {0,1, } R(i) = means that the variable x i is free (unrestricted)

29 Restric?ons f : {0,1} n {0,1} R : [n] {0,1, } Applying R to f, we get a Boolean func?on f R : {0,1} Stars(R) {0,1} R f R ( ) f ( )

30 Restric?ons Restric?ons are applied syntac3cally to DNFs / CNFs / decision trees / AC 0 circuits, etc.

31 Restric?ng an AC 0 Circuit Consider R = { x 1 1 } x 1 x 1 x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5

32 Restric?ng an AC 0 Circuit Consider R = { x 1 1 } i.e., R(1) = 1 and R(i) = for all i 1 x 1 x 1 x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5

33 Restric?ng an AC 0 Circuit Consider R = { x 1 1 } x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5

34 Restric?ng an AC 0 Circuit Consider R = { x 1 1 } x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5

35 Restric?ng an AC 0 Circuit Consider R = { x 1 1 } 1 0 x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5

36 Restric?ng an AC 0 Circuit Gate Elimina9on Method Consider R = { x 1 1 } N to 1+1/exp(d) lower bound against depth-d circuits via determinis3c restric3ons [Chaudhuri & Radhakrishnan 96] 1 0 x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5

37 Restric?ng a DNF R = { x 1 1, x 4 0 } x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 7 x 8 x 9 x 4 x 7 x 9

38 Restric?ng a DNF R = { x 1 1, x 4 0 } 1 x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 7 x 8 x 9 x 4 x 7 x 9

39 Restric?ng a DNF R = { x 1 1, x 4 0 } 1 x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 7 x 8 x 9 x 4 x 7 x 9

40 Restric?ng a DNF R = { x 1 1, x 4 0 } 1 x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 7 x 8 x 9 x 4 x 7 x 9

41 Restric?ng a DNF R = { x 1 1, x 4 0 } 1 x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 7 x 8 x 9 x 4 x 7 x 9 R x 2 x 3 x 2 x 5 x 7 x 8 x 9 x 7 x 9

42 Restric?ng a Decision Tree Again co R = { x 1 1 } x 3 x 1 x 7 x 4 x 7 x 1 x

43 Restric?ng a Decision Tree Again co R = { x 1 1 } x 3 x 1 x 7 x 4 x 7 x 1 x

44 Restric?ng a Decision Tree Again co R = { x 1 1 } x 3 R 0 x 3 1 x 1 x 7 x 7 x 7 x 4 x 7 x 1 x x 8 1 0

45 The Random Restric?on R p Throughout this talk, random objects in boldface

46 The Random Restric?on R p For 0 p 1, let R p be the random restric3on with R p (i) = with prob. p 0 with prob. (1-p)/2 1 with prob. (1-p)/2 independently for each variable index i [n]

47 Effect of R p R p simplifies Boolean func?ons represented by small decision trees / AC 0 circuits / demorgan formulas, [Subbotovskaya 61, ] Certain Boolean func?ons, like PARITY, maintain their complexity under R p => lower bounds!

48 Effect of R p on Boolean Func?ons R p kills AND n with very high probability (for p 1/2): Pr[ AND n R p 0 ] 1 (1/2) O(n)

49 Effect of R p on Boolean Func?ons R p kills AND n with very high probability (for p 1/2): Pr[ AND n R p 0 ] 1 (1/2) O(n) On the other hand, for any R with k stars, PARITY n R PARITY k or 1 PARITY k In par?cular, DT depth (PARITY n R p ) pn with high prob

50 Effect of R p on Decision Trees Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l

51 Effect of R p on Decision Trees Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l 2 l when p 1/4k

52 Effect of R p on DNFs Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l Hastad s Switching Lemma (1986) If F is a k-dnf (i.e. OR of depth-k decision trees), then Pr[ DT depth (F R p ) l ] (5pk) l

53 Effect of R p on DNFs Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t 2 l when R p ) p l ] 1/10k (2pk) l Hastad s Switching Lemma (1986) If F is a k-dnf (i.e. OR of depth-k decision trees), then Pr[ DT depth (F R p ) l ] (5pk) l

54 Effect of R p on DNFs Corollary (usual statement of the S.L.) If F is a k-dnf, then Pr[F R p is not equivalent to an l-cnf ] (5pk) l Hastad s Switching Lemma (1986) If F is a k-dnf (i.e. OR of depth-k decision trees), then Pr[ DT depth (F R p ) l ] (5pk) l

55 Effect of R p on DNFs CNFs Corollary (usual statement of the S.L.) If F is a k-cnf, then Pr[F R p is not equivalent to an l-dnf ] (5pk) l Hastad s Switching Lemma (1986) If F is a k-cnf (i.e. AND of depth-k decision trees), then Pr[ DT depth (F R p ) l ] (5pk) l

56 Effect of R p on DNFs k R p w.h.p. for small p l

57 Depth Reduc?on unbnd'd fan-in k k k k k k

58 Depth Reduc?on Apply the Switching Lemma to each gate (take a union bound over failure probability) k k k k k k

59 Depth Reduc?on Apply the Switching Lemma to each gate (take a union bound over failure probability) l l l l l l

60 Depth Reduc?on two layers of -gates l l l l l l

61 Depth Reduc?on l l l l l l

62 Depth Reduc?on Applying the S.L. to the boaom-2 layers k k k k k k l l l l l l

63 Depth Reduc?on Applying the S.L. to the boaom-2 layers k k k k k k typically, k = l n 1/d l l l l l l

64 Depth Reduc?on unbnd'd fan-in depth-k dec. trees

65 Hastad s Switching Lemma (decision tree version) If T 1,, T m are depth-k decision trees, then Pr[ DT depth ((T 1 T m ) R p ) l ] (5pk) l unbnd'd fan-in depth-k dec. trees

66 Hastad s Switching Lemma (decision tree version) If T 1,, T m are depth-k decision trees, then Pr[ DT depth ((T 1 T m ) R p ) l ] (5pk) l depth-l dec. trees

67 Depth Reduc?on applying the S.L. to the boaom layer of gates

68 PARITY Lower Bound

69 Hastad 86 Depth d+1 circuits for PARITY require size exp(ω(n 1/d ))

70 Hastad 86 Depth d+1 circuits for PARITY require size exp(ω(n 1/d )) (Trivial) Upper Bound PARITY has depth d+1 circuits of size exp(o(n 1/d )) d 2 n 1/d n 1 1/d n 1 1/d n 1 1/d

71 Hastad 86 Depth d+1 circuits for PARITY require size exp(ω(n 1/d )) (Trivial) Upper Bound PARITY has depth d+1 circuits of size exp(o(n 1/d )) d+1 n

72 PARITY Lower Bound x 1 x 1 x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5

73 PARITY Lower Bound depth-1 decision trees on n variables

74 PARITY Lower Bound R p -random restric?on ρ 1 : [n] {0,1, } depth-1 decision trees on n variables

75 PARITY Lower Bound Apply the Switching Lemma to each bo om-level gate Take a union bound over failure probabili?es depth-1 decision trees on n variables

76 PARITY Lower Bound Apply the Switching Lemma to each bo om-level gate Take a union bound over failure probabili?es low-depth decision trees on pn variables

77 PARITY Lower Bound 1 O(ε) frac?on have depth 0 O(ε) frac?on have depth 1 O(ε 2 ) frac?on have depth 2 O(ε l ) have depth l low-depth decision trees on pn variables

78 PARITY Lower Bound low-depth decision trees on pn variables

79 PARITY Lower Bound R p -random restric?on ρ 2 : Stars(ρ 1 ) {0,1, } low-depth decision trees on pn variables

80 PARITY Lower Bound low-depth decision trees on p 2 n variables

81 PARITY Lower Bound R p -random restric?on ρ 3 : Stars(ρ 2 ) {0,1, } low-depth decision trees on p 2 n variables

82 PARITY Lower Bound low-depth decision trees on p 3 n variables

83 PARITY Lower Bound R p -random restric?on ρ 4 : Stars(ρ 3 ) {0,1, } low-depth decision trees on p 3 n variables

84 PARITY Lower Bound depth-0 decision tree (i.e. constant funcgon) on p 4 n variables

85 PARITY Lower Bound depth-0 decision tree (i.e. constant funcgon) on p 4 n variables w.h.p., output is fixed to 0 or 1 a~er restric?ons ρ 1 ρ 2 ρ 3 ρ 4 (equivalent to R p 4) If p 4 n 1, this shows the original circuit does not compute PARITY n

86 PARITY Lower Bound Op?mal choice of parameters gives a?ght exp(ω(n 1/d )) lower bound for depth d+1 circuits x 1 x 1 x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5

87 Formulas vs. Circuits

88 AC 0 Circuits x 1 x 1 x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5

89 AC 0 Formulas x 3 x 7 x 4 x 1 x 3 x 2 x 11 x 1 x 2 x 5 x 4 x 2 x 1

90 Fact: Every depth d+1 circuit of size S is equivalent to a depth d+1 formula of size at most S d

91 PARITY Lower Bound Hastad 86 Depth d+1 circuits for PARITY require size exp(ω(n 1/d )) R. 15 Depth d+1 formulas for PARITY have size exp(ω(dn 1/d ))

92 PARITY Lower Bound Hastad 86 Depth d+1 circuits for PARITY require size exp(ω(n 1/d )) Poly-size circuits for PARITY have depth log n log log n + O(1) R. 15 Depth d+1 formulas for PARITY have size exp(ω(dn 1/d )) Poly-size formulas for PARITY have depth Ω(log n)

93 Two Views of R p : [n] {0,1, } (1) Independently, for each coordinate i [n], set R p (i) = with prob. p 0 with prob. (1-p)/2 1 with prob. (1-p)/2 (2) σ {0,1} n uniform random assignment iτ [0,1] n uniform random?me-stamp R p (i) = if τ i p σ i if τ i > p

94 1 τ 0 σ σ {0,1} n uniform random assignment iτ [0,1] n uniform random?me-stamp R p (i) = if τ i p σ i if τ i > p

95 1 τ?me p runs from 1 to 0 0 σ σ {0,1} n uniform random assignment iτ [0,1] n uniform random?me-stamp R p (i) = if τ i p σ i if τ i > p

96 1 start?me p = 1 τ 0 σ R p σ {0,1} n uniform random assignment iτ [0,1] n uniform random?me-stamp R p (i) = if τ i p σ i if τ i > p

97 1 start?me p = 1 τ 0 σ R p σ {0,1} n uniform random assignment iτ [0,1] n uniform random?me-stamp R p (i) = if τ i p σ i if τ i > p

98 1 p = 0.99 τ 0 σ R p 0 σ {0,1} n uniform random assignment iτ [0,1] n uniform random?me-stamp R p (i) = if τ i p σ i if τ i > p

99 1 p = 0.5 τ 0 σ R p σ {0,1} n uniform random assignment iτ [0,1] n uniform random?me-stamp R p (i) = if τ i p σ i if τ i > p

100 1 p = 0.1 τ 0 σ R p σ {0,1} n uniform random assignment iτ [0,1] n uniform random?me-stamp R p (i) = if τ i p σ i if τ i > p

101 1 end?me p = 0 τ 0 σ R p σ {0,1} n uniform random assignment iτ [0,1] n uniform random?me-stamp R p (i) = if τ i p σ i if τ i > p

102 1 0 x 1 x 1 x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5

103 1 p 0 R p x 1 x 1 x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5

104 1 p 0 p 2 R p 2 R p x 1 x 1 x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5

105 1 p 0 p 2 p 3 R p 3 R p 2 R p x 1 x 1 x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5

106 1 p 0 p 2 p 3 p 4 R p 4 R p 3 R p 2 R p x 1 x 1 x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5

107 1 0

108 1 0 Each sub-formula F will have its own stopping?me q(f) q(f) is a random variable in [0,1], which is determined by σ and τ

109 1 q(g) 0 G

110 1 q(h) 0 H

111 1 q(i) 0 I

112 1 q(j) 0 J

113 1 0 q(f) F

114 1 p 0

115 1 p 0 R p

116 1 p 0 decision trees

117 1 p 0 k

118 1 p 0 F k

119 1 p 0 q(f) F q(f) = p / 14k k

120 Hastad s Switching Lemma If F is k-dnf or k-cnf, then Pr[ DT depth (f R 1/14k ) l ] e l Stopping Time Version For every Boolean formula F, Pr[ DT depth (F R q(f) ) l ] e l

121 Technical Main Lemma (tail bound for q(f)) If F has depth d+1, then for all 0 λ 1, C Pr[ q(f) λ ] size(f) d exp(ω(dλ 1/d )) The constant C < 8 is defined by 1+ 1X i=0 1 1 exp(e i 1 (i + 1)e 2 ) + X i=0 1X j=0 1 exp((j + 1)e i 1 (i + j + 2)e 2 )

122 Technical Main Lemma (tail bound for q(f)) If F has depth d+1, then for all 0 λ 1, C Pr[ q(f) λ ] size(f) d exp(ω(dλ 1/d )) Corollary Depth d+1 formulas for PARITY have size exp(ω(dn 1/d ) O(d))

123 Proof of Hastad s Switching Lemma

124 Fix a k-dnf F and l 1 Switching Lemma: Pr[ depth(candt(f R p )) l ] = O(pk) l

125 Fix a k-dnf F and l 1 BAD := { restric?ons R depth(candt(f R)) l } Switching Lemma: Pr[ depth(candt(f R p )) l ] = O(pk) l

126 Fix a k-dnf F and l 1 BAD := { restric?ons R depth(candt(f R)) l } Goal. Pr[ R p BAD ] = O(pk) l Switching Lemma: Pr[ depth(candt(f R p )) l ] = O(pk) l

127 Fix a k-dnf F and l 1 BAD := { restric?ons R depth(candt(f R)) l } Goal. Pr[ R p BAD ] = O(pk) l Key idea. We associate each R BAD with a restric?on R* such that

128 Fix a k-dnf F and l 1 BAD := { restric?ons R depth(candt(f R)) l } Goal. Pr[ R p BAD ] = O(pk) l Key idea. We associate each R BAD with a restric?on R* such that 1 Stars(R*) = Stars(R) l R* R* (l = 3)

129 Fix a k-dnf F and l 1 BAD := { restric?ons R depth(candt(f R)) l } Goal. Pr[ R p BAD ] = O(pk) l Key idea. We associate each R BAD with a restric?on R* such that 1 Stars(R*) = Stars(R) l in par?cular, Pr[ R p = R* ] / Pr[ R p = R ] = ((1-p)/2p) l

130 Fix a k-dnf F and l 1 BAD := { restric?ons R depth(candt(f R)) l } Goal. Pr[ R p BAD ] = O(pk) l Key idea. We associate each R BAD with a restric?on R* such that 1 Stars(R*) = Stars(R) l in par?cular, Pr[ R p = R* ] / Pr[ R p = R ] = ((1-p)/2p) l for any restric?on ρ with s stars, Pr[ R p = ρ ] = p s (1 p) n s

131 Fix a k-dnf F and l 1 BAD := { restric?ons R depth(candt(f R)) l } Goal. Pr[ R p BAD ] = O(pk) l Key idea. We associate each R BAD with a restric?on R* such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1

132 Fix a k-dnf F and l 1 BAD := { restric?ons R depth(candt(f R)) l } Goal. Pr[ R p BAD ] = O(pk) l Key idea. We associate each R BAD with a restric?on R* such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1 any restric?on ρ equals R* for at most (4k) l dis?nct R BAD

133 Key idea. We associate each R BAD with a restric?on R* such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1

134 Key idea. We associate each R BAD with a restric?on R* such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1 Pr[ R p BAD ] = R BAD Pr[ R p = R ]

135 Key idea. We associate each R BAD with a restric?on R* such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1 Pr[ R p BAD ] = R BAD Pr[ R p = R ] = R BAD (2p/(1-p)) l Pr[ R p = R* ] Pr[ R p = R ] = (2p/(1-p)) l Pr[ R p = R* ]

136 Key idea. We associate each R BAD with a restric?on R* such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1 Pr[ R p BAD ] = R BAD Pr[ R p = R ] = R BAD (2p/(1-p)) l Pr[ R p = R* ] (4p) l R BAD Pr[ R p = R* ] w.l.o.g. p 1/2

137 Key idea. We associate each R BAD with a restric?on R* such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1 Pr[ R p BAD ] = R BAD Pr[ R p = R ] = R BAD (2p/(1-p)) l Pr[ R p = R* ] (4p) l R BAD Pr[ R p = R* ] (4p) l (4k) l Pr[ R p {R* R BAD} ]

138 Key idea. We associate each R BAD with a restric?on R* such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1 Pr[ R p BAD ] = R BAD Pr[ R p = R ] = R BAD (2p/(1-p)) l Pr[ R p = R* ] (4p) l R BAD Pr[ R p = R* ] (4p) l (4k) l Pr[ R p {R* R BAD} ] Pr[ ] 1

139 Key idea. We associate each R BAD with a restric?on R* such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1 Pr[ R p BAD ] = R BAD Pr[ R p = R ] = R BAD (2p/(1-p)) l Pr[ R p = R* ] (4p) l R BAD Pr[ R p = R* ] (4p) l (4k) l Pr[ R p {R* R BAD} ] (16pk) l Q.E.D.

140 Switching Lemma: Pr[ depth(candt(f R p )) l ] = O(pk) l Pr[ R p BAD ] = R BAD Pr[ R p = R ] = R BAD (2p/(1-p)) l Pr[ R p = R* ] (4p) l R BAD Pr[ R p = R* ] (4p) l (4k) l Pr[ R p {R* R BAD} ] (16pk) l Q.E.D.

141 Switching Lemma: Pr[ depth(candt(f R p )) l ] = O(pk) l Pr[ R p BAD ] = R BAD Pr[ R p = R ] = R BAD (2p/(1-p)) l Pr[ R p = R* ] more (4p) careful l R BAD analysis Pr[ R p = gives R* ] (5pk) l (w.l.o.g. p (4p) l (4k) l Pr[ R p {R* R BAD} ] (16pk) l Q.E.D.

142 Key idea. We associate each R BAD with a restric?on R* such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1

143 Key idea. We associate each R BAD with a pair (R*,Code(R)) such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1 3 4

144 Key idea. We associate each R BAD with a pair (R*,Code(R)) such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1 3 Code(R) ({0,1} 2 [k]) l 4

145 Key idea. We associate each R BAD with a pair (R*,Code(R)) such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1 3 Code(R) ({0,1} 2 [k]) l 4 in par?cular, Code(R) has (4k) l possible values

146 Key idea. We associate each R BAD with a pair (R*,Code(R)) such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1 3 Code(R) ({0,1} 2 [k]) l 4 the map R (R*,Code(R)) is 1-to-1

147 Key idea. We associate each R BAD with a pair (R*,Code(R)) such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1 3 Code(R) ({0,1} 2 [k]) l 4 the map R (R*,Code(R)) is 1-to-1 3 & 4 imply 2

148 Key idea. We associate each R BAD with a pair (R*,Code(R)) such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1 3 Code(R) ({0,1} 2 [k]) l 4 the map R (R*,Code(R)) is 1-to-1 intui?vely, Code(R) is a recipe for inver?ng R (R*,Code(R)) given knowledge of R*

149 R (R*,Code(R))

150 k = 3, l = 4 R (R*,Code(R)) F = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7

151 k = 3, l = 4 R (R*,Code(R)) F = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 }

152 k = 3, l = 4 R (R*,Code(R)) 1 1 F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 }

153 k = 3, l = 4 R (R*,Code(R)) 1 1 F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 }

154 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } x 2 CanDT(F R) = x 7 x 3 1 x 5 x

155 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R BAD, since CanDT(F R) l 4 x 2 x 7 x 3 1 x 5 x

156 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } x 2 x 7 x 3 1 x 5 x

157 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 } x 2 x 7 x 3 1 x 5 x

158 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 } x 2 x 7 x 3 1 x 5 x

159 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 } x 2 x 7 x 3 1 x 5 x

160 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 } x 2 x 7 x 3 1 x 5 x

161 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 } x 2 x 7 x 3 1 x 5 x

162 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 } Code(R) says: find the first sa?sfied term of F R* the long path begins with x 2 1, x 3 1 (i.e. var 2 of term 1 and var 3 of term 1)

163 k = 3, l = 4 R (R*,Code(R)) 1 F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } Fix variables according to the beginning of the long path R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 }

164 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 } x 2 x 7 x 3 1 x 5 x

165 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 } x 2 x 7 x 3 1 x 5 x

166 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 } x 2 x 7 x 3 1 x 5 x

167 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 } x 2 x 7 x 3 1 x 5 x

168 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 } Code(R) next says: find the next sa?sfied term of F R*(overwri?ng x 2 1, x 3 1) the long path con?nues x 5 0 (i.e. var 3 of term 0)

169 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } 0 R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } Fix variables according to the long path R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 }

170 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 } x 2 x 7 x 3 1 x 5 x

171 k = 3, l = 4 R (R*,Code(R)) F = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } Code(R) ({0,1} 2 [k]) l given knowledge of R* (and F), follow these instruc?ons to recover R (and along the way R~) R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 }

172 k = 3, l = 4 R (R*,Code(R)) F = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 } Code(R) ({0,1} 2 [k]) l given knowledge of R* (and F), follow these instruc?ons to recover R (and along the way R~) ü R* has l fewer stars than R ü R (R*,Code(R)) is 1-to-1

173 A Different Approach

174 A Different Approach Hastad s Switching Lemma If F is a k-dnf, then Pr[ depth(candt(f R p )) l ] = O(pk) l Weak (but flexible) Switching Lemma If F is a k-dnf, then Pr[ depth(candt(f) R p ) l ] = O(pk2 k ) l

175 Shows that PARITY has depth-d circuit size log n d 2 exp(ω(( ) )) Weak (but flexible) Switching Lemma If F is a k-dnf, then Pr[ depth(candt(f) R p ) l ] = O(pk2 k ) l

176 Proof generalizes to affine restric3ons! Weak (but flexible) Switching Lemma If F is a k-dnf, then Pr[ depth(candt(f) R p ) l ] = O(pk2 k ) l

177 A Different Approach 1. Decision Tree Switching Lemma 2. k-clipped Decision Trees 3. Arbitrary Distribu?on of Stars 4. Switching Lemma for Affine Restric3ons 5. Tsei?n Expander Switching Lemma

178 A Different Approach 1. Decision Tree Switching Lemma 2. k-clipped Decision Trees 3. Arbitrary Distribu?on of Stars 4. Switching Lemma for Affine Restric3ons 5. Tsei?n Expander Switching Lemma

179 Binomial Distribu?on For a finite set N, let S p N Denote S is a p-biased random subset of N, i.e., Pr[ S = S ] = p S (1 p) N S For a natural number n and 0 p 1, the binomial random variable Bin(n,p) is equivalent to S where S p [n]

180 Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l

181 Consider a random branch of T: x 3 x 1 x 2 1/4 x 4 x 7 x 7 1/8 x 2 1/16 1/16 1/8 1/8 1/8 1/8

182 Consider a random branch of T: x 3 x 1 x 2 1/4 x 4 x 7 x 7 1/8 x 2 1/16 1/16 1/8 1/8 1/8 1/8

183 Let β(t) [n] be the set of variables queried on a random branch of T x 3 x 1 x 2 1/4 x 4 x 7 x 7 1/8 x 2 1/16 1/16 1/8 1/8 1/8 1/8

184 #β(t) = the number of variables queried on a random branch of T x 3 x 1 x 2 1/4 x 4 x 7 x 7 1/8 x 2 1/16 1/16 1/8 1/8 1/8 1/8

185 #β(t) = the number of variables queried on a random branch of T x 3 Obs: If depth(t) k, then Pr[ #β(t) k ] 2 k x 1 x 2 1/4 x 4 x 7 x 7 1/8 x 2 1/16 1/16 1/8 1/8 1/8 1/8

186 Claim. The following random variables have the same distribu?on: 1. β(t R p ) the set of variables queried on a random branch of T R p 2. S p β(t) a p-biased subset of the vars. on a random branch of T

187 Claim. The following random variables have the same distribu?on: 1. β(t R p ) #β(t R p ) the set of variables queried on a random branch of T R p 2. S p β(t) Bin(#β(T), p) a p-biased subset of the vars. on a random branch of T

188 1. β(t R p ) 2. S p β(t) T

189 1. β(t R p ) 2. S p β(t) Stars(R p )

190 1. β(t R p ) 2. S p β(t) R p

191 1. β(t R p ) 2. S p β(t) T R p

192 1. β(t R p ) 2. S p β(t) T R p

193 1. β(t R p ) 2. S p β(t) T R p

194 1. β(t R p ) 2. S p β(t) T R p

195 1. β(t R p ) 2. S p β(t) β(t R p )

196 1. β(t R p ) 2. S p β(t) β(t R p )

197 1. β(t R p ) 2. S p β(t) β(t R p )

198 1. β(t R p ) 2. S p β(t) β(t R p )

199 1. β(t R p ) 2. S p β(t) β(t R p )

200 1. β(t R p ) 2. S p β(t) T

201 1. β(t R p ) 2. S p β(t) β(t)

202 1. β(t R p ) 2. S p β(t) β(t)

203 1. β(t R p ) 2. S p β(t) S p β(t)

204 1. β(t R p ) 2. S p β(t) 1. #β(t R p ) 2. Bin(β(T), p)

205 Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l

206 Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l Pr[ depth(t R p ) l ] recall: depth(t ) l Pr[ #β(t ) l ] 2 l

207 Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l Pr[ depth(t R p ) l ] Pr R p [ Pr β(t R p) [ #β(t R p ) l ] 2 l ] p p recall: depth(t ) l Pr[ #β(t ) l ] 2 l

208 Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l Pr[ depth(t R p ) l ] Pr R p [ Pr β(t R p) [ #β(t R p ) l ] 2 l ] p p

209 Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l Pr[ depth(t R p ) l ] Pr R p [ Pr β(t R p) [ #β(t R p ) l ] 2 l ] p 2 l Ex R p [ Pr β(t R p) [ #β(t R p ) l ] ] p p p Markov s inequality

210 Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l Pr[ depth(t R p ) l ] Pr R p [ Pr β(t R p) [ #β(t R p ) l ] 2 l ] p p 2 l Ex R p [ Pr β(t R p) [ #β(t R p ) l ] ] p p

211 Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l Pr[ depth(t R p ) l ] Pr R p [ Pr β(t R p) [ #β(t R p ) l ] 2 l ] p 2 l Pr[ #β(t R p ) l ] p

212 Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l Pr[ depth(t R p ) l ] Pr R p [ Pr β(t R p) [ #β(t R p ) l ] 2 l ] p 2 l Pr[ #β(t R p ) l ] = 2 l Pr[ Bin(#β(T), p) l ] p #β(t R p ) and Bin(#β(T), p) have the same distribu?on

213 Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l Pr[ depth(t R p ) l ] Pr R p [ Pr β(t R p) [ #β(t R p ) l ] 2 l ] p 2 l Pr[ #β(t R p ) l ] = 2 l Pr[ Bin(#β(T), p) l ] 2 l Pr[ Bin(k, p) l ] p #β(t) depth(t) k

214 Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l Pr[ depth(t R p ) l ] Pr R p [ Pr β(t R p) [ #β(t R p ) l ] 2 l ] p 2 l Pr[ #β(t R p ) l ] = 2 l Pr[ Bin(#β(T), p) l ] 2 l Pr[ Bin(k, p) l ] 2 l (pk) l p union bound

215 Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l Pr[ depth(t R p ) l ] Pr R p [ Pr β(t R p) [ #β(t R p ) l ] 2 l ] p 2 l Pr[ #β(t R p ) l ] = 2 l Pr[ Bin(#β(T), p) l ] 2 l Pr[ Bin(k, p) l ] 2 l (pk) l p Q.E.D.

216 A Different Approach 1. Decision Tree Switching Lemma 2. k-clipped Decision Trees 3. Arbitrary Distribu?on of Stars 4. Switching Lemma for Affine Restric3ons 5. Tsei?n Expander Switching Lemma

217 Recall: Canonical DT of a k-dnf (x 1 x 2 x 3 ) ( x 2 x 4 ) (x 1 x 3 x 7 ) x 4 1 x 2 1 x 1 x 3 x 7 x 4 x 2 1 x 7 1 x 3 1

218 Recall: Canonical DT of a k-dnf (x 1 x 2 x 3 ) ( x 2 x 4 ) (x 1 x 3 x 7 ) x 4 1 x 2 1 x 1 x 3 Obs: Every node has distance k from a leaf! x 7 x 4 x 2 1 x 7 1 x 3 1

219 Defini?on: A decision tree is k-clipped if every node has distance k from a leaf

220 Defini?on: A decision tree is k-clipped if every node has distance k from a leaf Lemma: If F is a k-dnf (or k-cnf), then CanDT(F) is k-clipped

221 Defini?on: A decision tree is k-clipped if every node has distance k from a leaf Lemma: If F is a k-dnf (or k-cnf), then CanDT(F) is k-clipped Lemma: If T is k-clipped, then Ex #β(t) l ( ) O(k2 k ) l

222 Defini?on: A decision tree is k-clipped if every node has distance k from a leaf #β(t) stochas?cally dominated by Lemma: If F is a k-dnf (or k-cnf), then the number of unbiased coin flips un?l CanDT(F) seeing is k consecu?ve k-clipped heads Lemma: If T is k-clipped, then Ex #β(t) l ( ) O(k2 k ) l

223 k-clipped Decision Tree Switching Lemma If T is a k-clipped decision tree, then Pr[ depth(t R p ) l ] = O(pk2 k ) l

224 k-clipped Decision Tree Switching Lemma If T is a k-clipped decision tree, then Pr[ depth(t R p ) l ] = O(pk2 k ) l Pr[ depth(t R p ) l ] Pr R p[ Pr β(t R p) [ #β(t R p ) l ] 2 l ] p 2 l Pr[ #β(t R p ) l ] p = 2 l Pr [ Bin(#β(T), p) l ] same as before

225 k-clipped Decision Tree Switching Lemma If T is a k-clipped decision tree, then Pr[ depth(t R p ) l ] = O(pk2 k ) l Pr[ depth(t R p ) l ] Pr R p[ Pr β(t R p) [ #β(t R p ) l ] 2 l ] p 2 l Pr[ #β(t R p ) l ] = 2 l Pr [ Bin(#β(T), p) l ] 2 l Ex[ p l ] p #β(t) l ( ) union bound

226 k-clipped Decision Tree Switching Lemma If T is a k-clipped decision tree, then Pr[ depth(t R p ) l ] = O(pk2 k ) l Lemma: If T is k-clipped, then Pr[ depth(t R p ) l ] Ex (#β(t)) Pr R p[ Pr β(t R p) [ #β(t O(k2 k R p ) ) l l l ] 2 l ] p 2 l Pr[ #β(t R p ) l ] = 2 l Pr [ Bin(#β(T), p) l ] 2 l Ex[ p l ] 2 l p l O(k2 k ) l p #β(t) l ( )

227 k-clipped Decision Tree Switching Lemma If T is a k-clipped decision tree, then Pr[ depth(t R p ) l ] = O(pk2 k ) l Pr[ depth(t R p ) l ] Pr R p[ Pr β(t R p) [ #β(t R p ) l ] 2 l ] p 2 l Pr[ #β(t R p ) l ] = 2 l Pr [ Bin(#β(T), p) l ] 2 l Ex[ p l ] 2 l p l O(k2 k ) l p #β(t) l ( ) Q.E.D.

228 A Different Approach 1. Decision Tree Switching Lemma 2. k-clipped Decision Trees 3. Arbitrary DistribuGon of Stars 4. Switching Lemma for Affine Restric3ons 5. Tsei?n Expander Switching Lemma

229 Let S [n] be an arbitrary distribu?on of stars

230 Let S [n] be an arbitrary distribu?on of stars Let u unif {0,1} n

231 Let S [n] be an arbitrary distribu?on of stars Let u unif {0,1} n Define random restric?on ρ S,u : [n] {0,1, } by ρ S,u (i) = S,u (i) = if i S, u i otherwise.

232 Let S [n] be an arbitrary distribu?on of stars Let u unif {0,1} n Define random restric?on ρ S,u : [n] {0,1, } by ρ S,u (i) = if i S, u i otherwise. If S p [n], then ρ S,u has the S,u has the same distribu?on as R p

233 Let S [n] be an arbitrary distribu?on of stars Defini?on: For p [0,1], we say that S is p-bounded if Pr[ J S ] p J for every J [n]

234 The following are equivalent: β(t) S 1. β(t ρ S,u ) 2. β(t) S

235 The following are equivalent: β(t) S 1. β(t ρ S,u ) 2. β(t) S in fact, this equivalence valid for fixed S Supp(S)

236 The following are equivalent: 1. β(t ρ S,u ) 2. β(t) S β(t) S If S is p-bounded, then Pr[ β(t) S l ] #β(t) l ( ) Ex[ p l ]

237 The following are equivalent: 1. β(t ρ S,u ) 2. β(t) S β(t) S Pr[ J S ] p l for every J [n] of size J = l If S is p-bounded, then Pr[ β(t) S l ] #β(t) l ( ) Ex[ p l ]

238 Theorem (Switching Lemma for ρ S,u ) If S is p-bounded and T is a k-clipped DT, then Pr[ depth(t ρ S,u ) l ] = O(pk2k k ) l

239 Theorem (Switching Lemma for ρ S,u ) If S is p-bounded and T is a k-clipped DT, then Pr[ depth(t ρ S,u ) l ] = O(pk2k k ) l Pr[ depth(t ρ S,u ) l ] Pr S,u [ Pr V(T ρ ) [ #β(t ρ S,u ) l ] 2 2 l l ] S,u 2 l Pr[ #β(t ρ S,u ) l ] = 2 l Pr [ β(t) S l ] #β(t) l ( ) 2 l Ex[ p l ] 2 l p l O(k2 k ) l Q.E.D.

240 A Different Approach 1. Decision Tree Switching Lemma 2. k-clipped Decision Trees 3. Arbitrary Distribu?on of Stars 4. Switching Lemma for Affine Restric9ons 5. Tsei?n Expander Switching Lemma

241 Restric?ons [n] {0,1, } are in 1-1 correspondence with subcubes of {0,1} n What about the complexity of F A (e.g. for F a k-dnf) for a random affine subspace of A {0,1} n?

242 Restric?ons [n] {0,1, } are in 1-1 correspondence with subcubes of {0,1} n What about the complexity of F A (e.g. for F a k-dnf) for a random affine subspace of A {0,1} n? A = V + u where V is a linear subspace of {0,1} n

243 Restric?ons [n] {0,1, } are in 1-1 correspondence with subcubes of {0,1} n What about the complexity of F A (e.g. for F a k-dnf) for a random affine subspace of A {0,1} n? Affine Switching Lemma If F is a k-dnf and A is a p-bounded random affine subset of {0,1} n, then Pr[ DT depth (F A) l ] = O(pk2 k ) l

244 Applying an Affine Restric?on Consider an affine set A {0,1} n For f : {0,1} n {0,1}, we define: f A : A {0,1}, DT depth (f A) = min. depth of a decision tree T such that f(x) = T(x) for all x A We also have a syntac3c opera?on T T A on decision trees

245 Applying an Affine Restric?on Example: A = { x {0,1} n x 1 x 3 } x 3 A 0 x 3 1 x 1 x 7 x 7 x 7 x 4 x 7 x 1 x x 8 1 0

246 Let S [n] be an arbitrary dist. of stars Let u unif {0,1} n Consider random restric?on ρ S,u (i) = S,u (i) = if i S, u i otherwise.

247 Let S [n] be an arbitrary dist. of stars Let V be an arbitrary dist. on linear subspaces of {0,1} n Let u unif {0,1} n Consider random restric?on ρ S,u (i) = S,u (i) = if i S, u i otherwise.

248 Let S [n] be an arbitrary dist. of stars Let V be an arbitrary dist. on linear subspaces of {0,1} n Let u unif {0,1} n Consider random restric?on ρ S,u (i) = Consider random affine space V + u if i S, u i otherwise.

249 Let S [n] be an arbitrary dist. of stars Let V be an arbitrary dist. on linear subspaces of {0,1} n Let u unif {0,1} n Consider random restric?on ρ S,u (i) = Consider random affine space V + u if i S, u i otherwise. V is the arrangement of stars

250 Let S [n] be an arbitrary dist. of stars Let V be an arbitrary dist. on linear subspaces of {0,1} n Let u unif {0,1} n Consider random restric?on ρ S,u (i) = Consider random affine space V + u if i S, u i otherwise. u gives the assignment to non-stars

251 Let S [n] be an arbitrary dist. of stars Let V be an arbitrary dist. on linear subspaces of {0,1} n Let u unif {0,1} n Consider random restric?on ρ S,u (i) = Consider random affine space V + u if i S, u i otherwise. {b 1,, b n } = standard basis for {0,1} n if V = Span{ b i i S }, then V + u is the subcube defined by ρ S,u

252 Defini?on A random set S [n] is p-bounded if, for all J [n], Pr[ J S ] p J Defini?on A random linear subspace V {0,1} n is p-bounded if, for all J [n], Pr[ V sha ers J ] p J

253 Defini?on A random set S [n] is p-bounded if, for all J [n], Pr[ J S ] p J Defini?on A random linear subspace V {0,1} n is p-bounded if, for all J [n], Pr[ V sha ers J ] p J I J v V such that v i = 1 for all i I v j = 0 for all j J I

254 Defini?on V sha ers J [n] A random set S [n] is p-bounded if, for all J [n], J is independent Pr[ J S ] p in J the matroid associated with V Defini?on A random linear subspace V {0,1} n is p-bounded if, for all J [n], Pr[ V sha ers J ] p J I J v V such that v i = 1 for all i I v j = 0 for all j J I

255 Defini?on A random set S [n] is p-bounded if, for all J [n], Pr[ J S ] p J Defini?on A random linear subspace V {0,1} n is p-bounded if, for all J [n], Pr[ V sha ers J ] p J equivalent if V = Span{ b i i S }

256 Suppose A = V + u where V is a p-bounded random linear subspace of {0,1} n u unif {0,1} n Theorem For every k-clipped decision tree T, Pr[ depth(t A) l ] = O(pk2 k ) l essen?ally the same proof as the Switching Lemma for ρ S,u

257 Suppose A = V + u where V is a p-bounded random linear subspace of {0,1} n u unif {0,1} n Theorem For every k-clipped decision tree T, Pr[ depth(t A) l ] = O(pk2 k ) l Corollary (Affine Switching Lemma) For every k-dnf F, Pr[ DT depth (F A) l ] = O(pk2 k ) l

258 Suppose A = V + u where V is a p-bounded random linear subspace of {0,1} n u unif {0,1} n Theorem For every k-clipped decision tree T, Pr[ depth(t A) l ] = O(pk2 k ) l Corollary (Affine Switching Lemma) For every k-dnf F, Pr[ DT depth (F A) l ] = O(pk2 k ) l Conjecture: O(pk) l

259 For every fixed V (and random u {0,1} n ), the following are equivalent: 1. β(t (V + u)) #β(t (V + u)) 2. greedy basis of β(t) in the V-matroid rank V (β(t))

260 Suppose A = V + u where V is a p-bounded random linear subspace of {0,1} n u unif {0,1} n Final generaliza?on: Replace the ambient space {0,1} n with an arbitrary affine space B

261 Suppose A = V + u where V is a p-bounded random linear subspace of B u unif B Final generaliza?on: Replace the ambient space {0,1} n with an arbitrary affine space B

262 Suppose A = V + u where V is a p-bounded random linear subspace of B u unif B A is a random affine subspace of B

263 Suppose A = V + u where V is a p-bounded random linear subspace of B u unif B Theorem For every B-independent k-clipped decision tree T, Pr[ depth(t A) l ] = O(pk2 k ) l

264 Suppose A = V + u where V is a p-bounded random linear subspace of B u unif B Theorem For every B-independent k-clipped decision tree T, Pr[ depth(t A) l ] = O(pk2 k ) l Every branch β [n] of T is sha ered by B (i.e. β is independent in the matroid associated with Lin(B) = { x y x,y B})

265 A Different Approach 1. Decision Tree Switching Lemma 2. k-clipped Decision Trees 3. Arbitrary Distribu?on of Stars 4. Switching Lemma for Affine Restric3ons 5. TseiGn Expander Switching Lemma

266 Tsei?n Expander Switching Lemma Theorem (Pitassi-R.-Servedio-Tan 16) Depth-d AC 0 Frege proofs of Tsei?n(3-regular log n d expander) require size exp(ω(( ) 2 )) Theorem (Beame-Pitassi-Impagliazzo 93, Ben-Sasson 02) Depth-d AC 0 Frege proofs of Tsei?n(3-regular expander) require size exp(ω(n 1/2 d ))

267 Tsei?n Expander Switching Lemma Theorem (Pitassi-R.-Servedio-Tan 16) Depth-d AC 0 Frege proofs of Tsei?n(3-regular log n d expander) require size exp(ω(( ) 2 )) Theorem (Beame-Pitassi-Impagliazzo 93, Ben-Sasson 02) Depth-d Improving AC 0 Frege the proofs AFFINE of SWITCHING Tsei?n(3-regular LEMMA would yield exp(ω(n 1/d expander) require size exp(ω(n 1/2 d )) ))

268 Suppose A = V + u where V is a p-bounded random linear subspace of {0,1} n u unif {0,1} n Affine Switching Lemma for k-dnfs For every k-dnf F, Pr[ DT depth (F A) l ] = O(pk2 k ) l Conjecture: O(pk) l

269 Stay Tuned Toni Pitassi: Expander S.L. & AC 0 -Frege lower bound Johan Hastad, Avishay Tal: Correla?on bounds from improved switching lemmas Johan Hastad, Rocco Servedio: Random projec?ons, depth hierarchy theorem Srikanth Srinivasan: Adap?ve random restric?ons (against AC 0 with few threshold gates)

270 Thanks!

CSC 2429 Approaches to the P vs. NP Question and Related Complexity Questions Lecture 2: Switching Lemma, AC 0 Circuit Lower Bounds

CSC 2429 Approaches to the P vs. NP Question and Related Complexity Questions Lecture 2: Switching Lemma, AC 0 Circuit Lower Bounds CSC 2429 Approaches to the P vs. NP Question and Related Complexity Questions Lecture 2: Switching Lemma, AC 0 Circuit Lower Bounds Lecturer: Toniann Pitassi Scribe: Robert Robere Winter 2014 1 Switching

More information

CS6840: Advanced Complexity Theory Mar 29, Lecturer: Jayalal Sarma M.N. Scribe: Dinesh K.

CS6840: Advanced Complexity Theory Mar 29, Lecturer: Jayalal Sarma M.N. Scribe: Dinesh K. CS684: Advanced Complexity Theory Mar 29, 22 Lecture 46 : Size lower bounds for AC circuits computing Parity Lecturer: Jayalal Sarma M.N. Scribe: Dinesh K. Theme: Circuit Complexity Lecture Plan: Proof

More information

CS 161: Design and Analysis of Algorithms

CS 161: Design and Analysis of Algorithms CS 161: Design and Analysis of Algorithms NP- Complete I P, NP Polynomial >me reduc>ons NP- Hard, NP- Complete Sat/ 3- Sat Decision Problem Suppose there is a func>on A that outputs True or False A decision

More information

Lecture 2: Proof of Switching Lemma

Lecture 2: Proof of Switching Lemma Lecture 2: oof of Switching Lemma Yuan Li June 25, 2014 1 Decision Tree Instead of bounding the probability that f can be written as some s-dnf, we estimate the probability that f can be computed by a

More information

CSCI 1010 Models of Computa3on. Lecture 02 Func3ons and Circuits

CSCI 1010 Models of Computa3on. Lecture 02 Func3ons and Circuits CSCI 1010 Models of Computa3on Lecture 02 Func3ons and Circuits Overview Func3ons and languages Designing circuits from func3ons Minterms and the DNF Maxterms and CNF Circuit complexity Algebra of Boolean

More information

Lecture 4: LMN Learning (Part 2)

Lecture 4: LMN Learning (Part 2) CS 294-114 Fine-Grained Compleity and Algorithms Sept 8, 2015 Lecture 4: LMN Learning (Part 2) Instructor: Russell Impagliazzo Scribe: Preetum Nakkiran 1 Overview Continuing from last lecture, we will

More information

Lecture 5: February 21, 2017

Lecture 5: February 21, 2017 COMS 6998: Advanced Complexity Spring 2017 Lecture 5: February 21, 2017 Lecturer: Rocco Servedio Scribe: Diana Yin 1 Introduction 1.1 Last Time 1. Established the exact correspondence between communication

More information

On the correlation of parity and small-depth circuits

On the correlation of parity and small-depth circuits Electronic Colloquium on Computational Complexity, Report No. 137 (2012) On the correlation of parity and small-depth circuits Johan Håstad KTH - Royal Institute of Technology November 1, 2012 Abstract

More information

Lecture 10: Learning DNF, AC 0, Juntas. 1 Learning DNF in Almost Polynomial Time

Lecture 10: Learning DNF, AC 0, Juntas. 1 Learning DNF in Almost Polynomial Time Analysis of Boolean Functions (CMU 8-859S, Spring 2007) Lecture 0: Learning DNF, AC 0, Juntas Feb 5, 2007 Lecturer: Ryan O Donnell Scribe: Elaine Shi Learning DNF in Almost Polynomial Time From previous

More information

Guest Lecture: Average Case depth hierarchy for,, -Boolean circuits. Boaz Barak

Guest Lecture: Average Case depth hierarchy for,, -Boolean circuits. Boaz Barak Guest Lecture: Average Case depth hierarchy for,, -Boolean circuits. Boaz Barak Thanks to the authors of RST for answering my questions by email and in particular to Li-Yang Tan for sharing with me his

More information

Testing by Implicit Learning

Testing by Implicit Learning Testing by Implicit Learning Rocco Servedio Columbia University ITCS Property Testing Workshop Beijing January 2010 What this talk is about 1. Testing by Implicit Learning: method for testing classes of

More information

complexity distributions

complexity distributions The of complexity distributions Emanuele Viola Northeastern University March 2012 Local functions (a.k.a. Junta, NC 0 ) f : {0,1}n {0,1} d-local : output depends on d input bits Input x d f Fact: Parity(x)

More information

Lecture 5: February 16, 2012

Lecture 5: February 16, 2012 COMS 6253: Advanced Computational Learning Theory Lecturer: Rocco Servedio Lecture 5: February 16, 2012 Spring 2012 Scribe: Igor Carboni Oliveira 1 Last time and today Previously: Finished first unit on

More information

1 AC 0 and Håstad Switching Lemma

1 AC 0 and Håstad Switching Lemma princeton university cos 522: computational complexity Lecture 19: Circuit complexity Lecturer: Sanjeev Arora Scribe:Tony Wirth Complexity theory s Waterloo As we saw in an earlier lecture, if PH Σ P 2

More information

Learning and Fourier Analysis

Learning and Fourier Analysis Learning and Fourier Analysis Grigory Yaroslavtsev http://grigory.us Slides at http://grigory.us/cis625/lecture2.pdf CIS 625: Computational Learning Theory Fourier Analysis and Learning Powerful tool for

More information

Subspace-Invariant AC 0 Formulas

Subspace-Invariant AC 0 Formulas Subspace-Invariant AC 0 Formulas Benjamin Rossman University of Toronto June 12, 2018 Abstract We consider the action of a linear subspace U of {0, 1} n on the set of AC 0 formulas with inputs labeled

More information

Lecture 3: AC 0, the switching lemma

Lecture 3: AC 0, the switching lemma Lecture 3: AC 0, the switching lemma Topics in Complexity Theory and Pseudorandomness (Spring 2013) Rutgers University Swastik Kopparty Scribes: Meng Li, Abdul Basit 1 Pseudorandom sets We start by proving

More information

3 Finish learning monotone Boolean functions

3 Finish learning monotone Boolean functions COMS 6998-3: Sub-Linear Algorithms in Learning and Testing Lecturer: Rocco Servedio Lecture 5: 02/19/2014 Spring 2014 Scribes: Dimitris Paidarakis 1 Last time Finished KM algorithm; Applications of KM

More information

CS Communication Complexity: Applications and New Directions

CS Communication Complexity: Applications and New Directions CS 2429 - Communication Complexity: Applications and New Directions Lecturer: Toniann Pitassi 1 Introduction In this course we will define the basic two-party model of communication, as introduced in the

More information

6.842 Randomness and Computation April 2, Lecture 14

6.842 Randomness and Computation April 2, Lecture 14 6.84 Randomness and Computation April, 0 Lecture 4 Lecturer: Ronitt Rubinfeld Scribe: Aaron Sidford Review In the last class we saw an algorithm to learn a function where very little of the Fourier coeffecient

More information

Learning Sums of Independent Integer Random Variables

Learning Sums of Independent Integer Random Variables Learning Sums of Independent Integer Random Variables Cos8s Daskalakis (MIT) Ilias Diakonikolas (Edinburgh) Ryan O Donnell (CMU) Rocco Servedio (Columbia) Li- Yang Tan (Columbia) FOCS 2013, Berkeley CA

More information

Two Sides of the Coin Problem

Two Sides of the Coin Problem Electronic Colloquium on Computational Complexity, Report No. 23 (2014) Two Sides of the Coin Problem Gil Cohen Anat Ganor Ran Raz February 19, 2014 Abstract In the coin problem, one is given n independent

More information

Notes for Lecture 25

Notes for Lecture 25 U.C. Berkeley CS278: Computational Complexity Handout N25 ofessor Luca Trevisan 12/1/2004 Notes for Lecture 25 Circuit Lower Bounds for Parity Using Polynomials In this lecture we prove a lower bound on

More information

Some Depth Two (and Three) Threshold Circuit Lower Bounds. Ryan Williams Stanford Joint work with Daniel Kane (UCSD)

Some Depth Two (and Three) Threshold Circuit Lower Bounds. Ryan Williams Stanford Joint work with Daniel Kane (UCSD) Some Depth Two (and Three) Threshold Circuit Lower Bounds Ryan Williams Stanford Joint wor with Daniel Kane (UCSD) Introduction Def. f n : 0,1 n 0,1 is a linear threshold function (LTF) if there are w

More information

Linear sketching for Functions over Boolean Hypercube

Linear sketching for Functions over Boolean Hypercube Linear sketching for Functions over Boolean Hypercube Grigory Yaroslavtsev (Indiana University, Bloomington) http://grigory.us with Sampath Kannan (U. Pennsylvania), Elchanan Mossel (MIT) and Swagato Sanyal

More information

Certifying polynomials for AC 0 [ ] circuits, with applications

Certifying polynomials for AC 0 [ ] circuits, with applications Certifying polynomials for AC 0 [ ] circuits, with applications Swastik Kopparty Srikanth Srinivasan Abstract In this paper, we introduce and develop the method of certifying polynomials for proving AC

More information

CS Introduction to Complexity Theory. Lecture #11: Dec 8th, 2015

CS Introduction to Complexity Theory. Lecture #11: Dec 8th, 2015 CS 2401 - Introduction to Complexity Theory Lecture #11: Dec 8th, 2015 Lecturer: Toniann Pitassi Scribe Notes by: Xu Zhao 1 Communication Complexity Applications Communication Complexity (CC) has many

More information

Learning and Fourier Analysis

Learning and Fourier Analysis Learning and Fourier Analysis Grigory Yaroslavtsev http://grigory.us CIS 625: Computational Learning Theory Fourier Analysis and Learning Powerful tool for PAC-style learning under uniform distribution

More information

Complexity Theory of Polynomial-Time Problems

Complexity Theory of Polynomial-Time Problems Complexity Theory of Polynomial-Time Problems Lecture 3: The polynomial method Part I: Orthogonal Vectors Sebastian Krinninger Organization of lecture No lecture on 26.05. (State holiday) 2 nd exercise

More information

Testing Monotone High-Dimensional Distributions

Testing Monotone High-Dimensional Distributions Testing Monotone High-Dimensional Distributions Ronitt Rubinfeld Computer Science & Artificial Intelligence Lab. MIT Cambridge, MA 02139 ronitt@theory.lcs.mit.edu Rocco A. Servedio Department of Computer

More information

Randomized Complexity Classes; RP

Randomized Complexity Classes; RP Randomized Complexity Classes; RP Let N be a polynomial-time precise NTM that runs in time p(n) and has 2 nondeterministic choices at each step. N is a polynomial Monte Carlo Turing machine for a language

More information

Tight bounds on The Fourier Spectrum of AC 0

Tight bounds on The Fourier Spectrum of AC 0 Electronic Colloquium on Computational Complexity, Report No. 174 (2014) Tight bounds on The Fourier Spectrum of AC 0 Avishay Tal December 14, 2014 Abstract We show that AC 0 circuits of depth d and size

More information

5.1 Learning using Polynomial Threshold Functions

5.1 Learning using Polynomial Threshold Functions CS 395T Computational Learning Theory Lecture 5: September 17, 2007 Lecturer: Adam Klivans Scribe: Aparajit Raghavan 5.1 Learning using Polynomial Threshold Functions 5.1.1 Recap Definition 1 A function

More information

arxiv: v1 [cs.cc] 13 Feb 2017

arxiv: v1 [cs.cc] 13 Feb 2017 Separation of AC 0 [ ] Formulas and Circuits arxiv:1702.03625v1 [cs.cc] 13 Feb 2017 Benjamin Rossman Depts. of Math and Computer Science University of Toronto February 14, 2017 Abstract Srikanth Srinivasan

More information

Lecture 29: Computational Learning Theory

Lecture 29: Computational Learning Theory CS 710: Complexity Theory 5/4/2010 Lecture 29: Computational Learning Theory Instructor: Dieter van Melkebeek Scribe: Dmitri Svetlov and Jake Rosin Today we will provide a brief introduction to computational

More information

CS 395T Computational Learning Theory. Scribe: Mike Halcrow. x 4. x 2. x 6

CS 395T Computational Learning Theory. Scribe: Mike Halcrow. x 4. x 2. x 6 CS 395T Computational Learning Theory Lecture 3: September 0, 2007 Lecturer: Adam Klivans Scribe: Mike Halcrow 3. Decision List Recap In the last class, we determined that, when learning a t-decision list,

More information

Communication is bounded by root of rank

Communication is bounded by root of rank Electronic Colloquium on Computational Complexity, Report No. 84 (2013) Communication is bounded by root of rank Shachar Lovett June 7, 2013 Abstract We prove that any total boolean function of rank r

More information

Notes for Lecture 11

Notes for Lecture 11 Stanford University CS254: Computational Complexity Notes 11 Luca Trevisan 2/11/2014 Notes for Lecture 11 Circuit Lower Bounds for Parity Using Polynomials In this lecture we prove a lower bound on the

More information

1 Randomized reduction: a first start

1 Randomized reduction: a first start 6.S078 Fine-Grained Algorithms and Complexity MIT Lectures 3 and 4 February 14 and 21, 2018 Today: In the previous two lectures, we mentioned ETH and SETH, about the time complexity of SAT. Today we want

More information

Meta-Algorithms vs. Circuit Lower Bounds Valentine Kabanets

Meta-Algorithms vs. Circuit Lower Bounds Valentine Kabanets Meta-Algorithms vs. Circuit Lower Bounds Valentine Kabanets Tokyo Institute of Technology & Simon Fraser University Understanding Efficiency Better understanding of Efficient Computation Good Algorithms

More information

An Improved Deterministic #SAT Algorithm for Small De Morgan Formulas

An Improved Deterministic #SAT Algorithm for Small De Morgan Formulas Noname manuscript No. (will be inserted by the editor) An Improved Deterministic #SAT Algorithm for Small De Morgan Formulas Ruiwen Chen Valentine Kabanets Nitin Saurabh Received: date / Accepted: date

More information

c 2004 Society for Industrial and Applied Mathematics

c 2004 Society for Industrial and Applied Mathematics SIAM J. COMPUT. Vol. 34, No. 2, pp. 261 276 c 2004 Society for Industrial and Applied Mathematics BOUNDED-DEPTH FREGE LOWER BOUNDS FOR WEAKER PIGEONHOLE PRINCIPLES JOSHUA BURESH-OPPENHEIM, PAUL BEAME,

More information

On converting CNF to DNF

On converting CNF to DNF Electronic Colloquium on Computational Complexity, Report No. 17 (2003) On converting CNF to DNF Peter Bro Miltersen Jaikumar Radhakrishnan Ingo Wegener March 27, 2003 Abstract The best-known representations

More information

Multi-Linear Formulas for Permanent and Determinant are of Super-Polynomial Size

Multi-Linear Formulas for Permanent and Determinant are of Super-Polynomial Size Multi-Linear Formulas for Permanent and Determinant are of Super-Polynomial Size Ran Raz Weizmann Institute ranraz@wisdom.weizmann.ac.il Abstract An arithmetic formula is multi-linear if the polynomial

More information

Complete problems for classes in PH, The Polynomial-Time Hierarchy (PH) oracle is like a subroutine, or function in

Complete problems for classes in PH, The Polynomial-Time Hierarchy (PH) oracle is like a subroutine, or function in Oracle Turing Machines Nondeterministic OTM defined in the same way (transition relation, rather than function) oracle is like a subroutine, or function in your favorite PL but each call counts as single

More information

The Analysis of Partially Symmetric Functions

The Analysis of Partially Symmetric Functions The Analysis of Partially Symmetric Functions Eric Blais Based on joint work with Amit Weinstein Yuichi Yoshida Classes of simple functions Classes of simple functions Constant Classes of simple functions

More information

CS Foundations of Communication Complexity

CS Foundations of Communication Complexity CS 49 - Foundations of Communication Complexity Lecturer: Toniann Pitassi 1 The Discrepancy Method Cont d In the previous lecture we ve outlined the discrepancy method, which is a method for getting lower

More information

Lecture 3: Nondeterminism, NP, and NP-completeness

Lecture 3: Nondeterminism, NP, and NP-completeness CSE 531: Computational Complexity I Winter 2016 Lecture 3: Nondeterminism, NP, and NP-completeness January 13, 2016 Lecturer: Paul Beame Scribe: Paul Beame 1 Nondeterminism and NP Recall the definition

More information

Structure of protocols for XOR functions

Structure of protocols for XOR functions Electronic Colloquium on Computational Complexity, Report No. 44 (016) Structure of protocols for XOR functions Kaave Hosseini Computer Science and Engineering University of California, San Diego skhossei@ucsd.edu

More information

Lecture Notes 3 Convergence (Chapter 5)

Lecture Notes 3 Convergence (Chapter 5) Lecture Notes 3 Convergence (Chapter 5) 1 Convergence of Random Variables Let X 1, X 2,... be a sequence of random variables and let X be another random variable. Let F n denote the cdf of X n and let

More information

CMPUT 675: Approximation Algorithms Fall 2014

CMPUT 675: Approximation Algorithms Fall 2014 CMPUT 675: Approximation Algorithms Fall 204 Lecture 25 (Nov 3 & 5): Group Steiner Tree Lecturer: Zachary Friggstad Scribe: Zachary Friggstad 25. Group Steiner Tree In this problem, we are given a graph

More information

Upper Bounds on Fourier Entropy

Upper Bounds on Fourier Entropy Upper Bounds on Fourier Entropy Sourav Chakraborty 1, Raghav Kulkarni 2, Satyanarayana V. Lokam 3, and Nitin Saurabh 4 1 Chennai Mathematical Institute, Chennai, India sourav@cmi.ac.in 2 Centre for Quantum

More information

The Strength of Multilinear Proofs

The Strength of Multilinear Proofs The Strength of Multilinear Proofs Ran Raz Iddo Tzameret December 19, 2006 Abstract We introduce an algebraic proof system that manipulates multilinear arithmetic formulas. We show that this proof system

More information

Regular Resolution Lower Bounds for the Weak Pigeonhole Principle

Regular Resolution Lower Bounds for the Weak Pigeonhole Principle Regular Resolution Lower Bounds for the Weak Pigeonhole Principle Toniann Pitassi Department of Computer Science University of Toronto toni@cs.toronto.edu Ran Raz Department of Computer Science Weizmann

More information

Improved Direct Product Theorems for Randomized Query Complexity

Improved Direct Product Theorems for Randomized Query Complexity Improved Direct Product Theorems for Randomized Query Complexity Andrew Drucker Nov. 16, 2010 Andrew Drucker, Improved Direct Product Theorems for Randomized Query Complexity 1/28 Big picture Usually,

More information

Junta Approximations for Submodular, XOS and Self-Bounding Functions

Junta Approximations for Submodular, XOS and Self-Bounding Functions Junta Approximations for Submodular, XOS and Self-Bounding Functions Vitaly Feldman Jan Vondrák IBM Almaden Research Center Simons Institute, Berkeley, October 2013 Feldman-Vondrák Approximations by Juntas

More information

CS 151 Complexity Theory Spring Solution Set 5

CS 151 Complexity Theory Spring Solution Set 5 CS 151 Complexity Theory Spring 2017 Solution Set 5 Posted: May 17 Chris Umans 1. We are given a Boolean circuit C on n variables x 1, x 2,..., x n with m, and gates. Our 3-CNF formula will have m auxiliary

More information

Average-Case Lower Bounds and Satisfiability Algorithms for Small Threshold Circuits

Average-Case Lower Bounds and Satisfiability Algorithms for Small Threshold Circuits Average-Case Lower Bounds and Satisfiability Algorithms for Small Threshold Circuits Ruiwen Chen 1, Rahul Santhanam 2, and Srikanth Srinivasan 3 1 University of Oxford, Oxford, UK ruiwen.chen@cs.ox.ac.uk

More information

Efficient Approximation for Restricted Biclique Cover Problems

Efficient Approximation for Restricted Biclique Cover Problems algorithms Article Efficient Approximation for Restricted Biclique Cover Problems Alessandro Epasto 1, *, and Eli Upfal 2 ID 1 Google Research, New York, NY 10011, USA 2 Department of Computer Science,

More information

Graphical Models. Lecture 10: Variable Elimina:on, con:nued. Andrew McCallum

Graphical Models. Lecture 10: Variable Elimina:on, con:nued. Andrew McCallum Graphical Models Lecture 10: Variable Elimina:on, con:nued Andrew McCallum mccallum@cs.umass.edu Thanks to Noah Smith and Carlos Guestrin for some slide materials. 1 Last Time Probabilis:c inference is

More information

Evaluation of DNF Formulas

Evaluation of DNF Formulas Evaluation of DNF Formulas Sarah R. Allen Carnegie Mellon University Computer Science Department Pittsburgh, PA, USA Lisa Hellerstein and Devorah Kletenik Polytechnic Institute of NYU Dept. of Computer

More information

Alternating nonzero automata

Alternating nonzero automata Alternating nonzero automata Application to the satisfiability of CTL [,, P >0, P =1 ] Hugo Gimbert, joint work with Paulin Fournier LaBRI, Université de Bordeaux ANR Stoch-MC 06/07/2017 Control and verification

More information

ASwitchingLemmaPrimer

ASwitchingLemmaPrimer ASwitchingLemmaPrimer Paul Beame University of Toronto 10 King s College Road Toronto, Ontario Canada M5S 1A4 beame@cs.washington.edu April 94 Abstract We present simplified proofs for a variety of switching

More information

IST 4 Information and Logic

IST 4 Information and Logic IST 4 Information and Logic mon tue wed thr fri sun T = today 3 M oh x= hw#x out 0 oh M 7 oh oh 2 M2 oh oh x= hw#x due 24 oh oh 2 oh = office hours oh oh M2 8 3 oh midterms oh oh Mx= MQx out 5 oh 3 4 oh

More information

Expectation. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Expectation. DS GA 1002 Statistical and Mathematical Models.   Carlos Fernandez-Granda Expectation DS GA 1002 Statistical and Mathematical Models http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall16 Carlos Fernandez-Granda Aim Describe random variables with a few numbers: mean, variance,

More information

On the Fourier spectrum of symmetric Boolean functions

On the Fourier spectrum of symmetric Boolean functions On the Fourier spectrum of symmetric Boolean functions Amir Shpilka Technion and MSR NE Based on joint work with Avishay Tal 1 Theme: Analysis of Boolean functions Pick favorite representation: Fourier

More information

Poly-logarithmic independence fools AC 0 circuits

Poly-logarithmic independence fools AC 0 circuits Poly-logarithmic independence fools AC 0 circuits Mark Braverman Microsoft Research New England January 30, 2009 Abstract We prove that poly-sized AC 0 circuits cannot distinguish a poly-logarithmically

More information

The Monte Carlo Method

The Monte Carlo Method The Monte Carlo Method Example: estimate the value of π. Choose X and Y independently and uniformly at random in [0, 1]. Let Pr(Z = 1) = π 4. 4E[Z] = π. { 1 if X Z = 2 + Y 2 1, 0 otherwise, Let Z 1,...,

More information

Lecture 24: Randomized Complexity, Course Summary

Lecture 24: Randomized Complexity, Course Summary 6.045 Lecture 24: Randomized Complexity, Course Summary 1 1/4 1/16 1/4 1/4 1/32 1/16 1/32 Probabilistic TMs 1/16 A probabilistic TM M is a nondeterministic TM where: Each nondeterministic step is called

More information

Evolutionary Dynamics on Graphs

Evolutionary Dynamics on Graphs Evolutionary Dynamics on Graphs Leslie Ann Goldberg, University of Oxford Absorption Time of the Moran Process (2014) with Josep Díaz, David Richerby and Maria Serna Approximating Fixation Probabilities

More information

Low degree almost Boolean functions are sparse juntas

Low degree almost Boolean functions are sparse juntas Low degree almost Boolean functions are sparse juntas Irit Dinur Yuval Filmus Prahladh Harsha November 19, 2017 Abstract Nisan and Szegedy showed that low degree Boolean functions are juntas. Kindler and

More information

Lower Bounds for Cutting Planes Proofs. with Small Coecients. Abstract. We consider small-weight Cutting Planes (CP ) proofs; that is,

Lower Bounds for Cutting Planes Proofs. with Small Coecients. Abstract. We consider small-weight Cutting Planes (CP ) proofs; that is, Lower Bounds for Cutting Planes Proofs with Small Coecients Maria Bonet y Toniann Pitassi z Ran Raz x Abstract We consider small-weight Cutting Planes (CP ) proofs; that is, Cutting Planes (CP ) proofs

More information

Lecture 16: Communication Complexity

Lecture 16: Communication Complexity CSE 531: Computational Complexity I Winter 2016 Lecture 16: Communication Complexity Mar 2, 2016 Lecturer: Paul Beame Scribe: Paul Beame 1 Communication Complexity In (two-party) communication complexity

More information

Quantum algorithms based on span programs

Quantum algorithms based on span programs Quantum algorithms based on span programs Ben Reichardt IQC, U Waterloo [arxiv:0904.2759] Model: Complexity measure: Quantum algorithms Black-box query complexity Span programs Witness size [KW 93] [RŠ

More information

Resolution Lower Bounds for the Weak Pigeonhole Principle

Resolution Lower Bounds for the Weak Pigeonhole Principle Electronic Colloquium on Computational Complexity, Report No. 21 (2001) Resolution Lower Bounds for the Weak Pigeonhole Principle Ran Raz Weizmann Institute, and The Institute for Advanced Study ranraz@wisdom.weizmann.ac.il

More information

Clique is hard on average for regular resolution

Clique is hard on average for regular resolution Clique is hard on average for regular resolution Ilario Bonacina, UPC Barcelona Tech July 20, 2018 RaTLoCC, Bertinoro How hard is to certify that a graph is Ramsey? Ilario Bonacina, UPC Barcelona Tech

More information

FPT hardness for Clique and Set Cover with super exponential time in k

FPT hardness for Clique and Set Cover with super exponential time in k FPT hardness for Clique and Set Cover with super exponential time in k Mohammad T. Hajiaghayi Rohit Khandekar Guy Kortsarz Abstract We give FPT-hardness for setcover and clique with super exponential time

More information

Majority is incompressible by AC 0 [p] circuits

Majority is incompressible by AC 0 [p] circuits Majority is incompressible by AC 0 [p] circuits Igor Carboni Oliveira Columbia University Joint work with Rahul Santhanam (Univ. Edinburgh) 1 Part 1 Background, Examples, and Motivation 2 Basic Definitions

More information

Active Testing! Liu Yang! Joint work with Nina Balcan,! Eric Blais and Avrim Blum! Liu Yang 2011

Active Testing! Liu Yang! Joint work with Nina Balcan,! Eric Blais and Avrim Blum! Liu Yang 2011 ! Active Testing! Liu Yang! Joint work with Nina Balcan,! Eric Blais and Avrim Blum! 1 Property Testing Instance space X = R n (Distri D over X)! Tested function f : X->{0,1}! A property P of Boolean fn

More information

Lecture 4. 1 Circuit Complexity. Notes on Complexity Theory: Fall 2005 Last updated: September, Jonathan Katz

Lecture 4. 1 Circuit Complexity. Notes on Complexity Theory: Fall 2005 Last updated: September, Jonathan Katz Notes on Complexity Theory: Fall 2005 Last updated: September, 2005 Jonathan Katz Lecture 4 1 Circuit Complexity Circuits are directed, acyclic graphs where nodes are called gates and edges are called

More information

A Noisy-Influence Regularity Lemma for Boolean Functions Chris Jones

A Noisy-Influence Regularity Lemma for Boolean Functions Chris Jones A Noisy-Influence Regularity Lemma for Boolean Functions Chris Jones Abstract We present a regularity lemma for Boolean functions f : {, } n {, } based on noisy influence, a measure of how locally correlated

More information

COMP 562: Introduction to Machine Learning

COMP 562: Introduction to Machine Learning COMP 562: Introduction to Machine Learning Lecture 20 : Support Vector Machines, Kernels Mahmoud Mostapha 1 Department of Computer Science University of North Carolina at Chapel Hill mahmoudm@cs.unc.edu

More information

Distributed Systems Gossip Algorithms

Distributed Systems Gossip Algorithms Distributed Systems Gossip Algorithms He Sun School of Informatics University of Edinburgh What is Gossip? Gossip algorithms In a gossip algorithm, each node in the network periodically exchanges information

More information

Approximation Algorithms and Hardness of Approximation. IPM, Jan Mohammad R. Salavatipour Department of Computing Science University of Alberta

Approximation Algorithms and Hardness of Approximation. IPM, Jan Mohammad R. Salavatipour Department of Computing Science University of Alberta Approximation Algorithms and Hardness of Approximation IPM, Jan 2006 Mohammad R. Salavatipour Department of Computing Science University of Alberta 1 Introduction For NP-hard optimization problems, we

More information

6.895 Randomness and Computation March 19, Lecture Last Lecture: Boosting Weak Learners Into Strong Learners

6.895 Randomness and Computation March 19, Lecture Last Lecture: Boosting Weak Learners Into Strong Learners 6.895 Randomness and Computation March 9, 2008 Lecture 3 Lecturer: Ronitt Rubinfeld Scribe: Edwin Chen Overview. Last Lecture: Boosting Weak Learners Into Strong Learners In the last two lectures, we showed

More information

CSC 5170: Theory of Computational Complexity Lecture 9 The Chinese University of Hong Kong 15 March 2010

CSC 5170: Theory of Computational Complexity Lecture 9 The Chinese University of Hong Kong 15 March 2010 CSC 5170: Theory of Computational Complexity Lecture 9 The Chinese University of Hong Kong 15 March 2010 We now embark on a study of computational classes that are more general than NP. As these classes

More information

Circuits. Lecture 11 Uniform Circuit Complexity

Circuits. Lecture 11 Uniform Circuit Complexity Circuits Lecture 11 Uniform Circuit Complexity 1 Recall 2 Recall Non-uniform complexity 2 Recall Non-uniform complexity P/1 Decidable 2 Recall Non-uniform complexity P/1 Decidable NP P/log NP = P 2 Recall

More information

ACO Comprehensive Exam March 20 and 21, Computability, Complexity and Algorithms

ACO Comprehensive Exam March 20 and 21, Computability, Complexity and Algorithms 1. Computability, Complexity and Algorithms Part a: You are given a graph G = (V,E) with edge weights w(e) > 0 for e E. You are also given a minimum cost spanning tree (MST) T. For one particular edge

More information

Machine Learning

Machine Learning Machine Learning 10-601 Tom M. Mitchell Machine Learning Department Carnegie Mellon University August 30, 2017 Today: Decision trees Overfitting The Big Picture Coming soon Probabilistic learning MLE,

More information

Binary Decision Diagrams Boolean Functions

Binary Decision Diagrams Boolean Functions Binary Decision Diagrams Representation of Boolean Functions BDDs, OBDDs, ROBDDs Operations Model-Checking over BDDs 72 Boolean functions:b = {0,1}, f :B B B Boolean Functions Boolean expressions: t ::=

More information

Polynomial time Prediction Strategy with almost Optimal Mistake Probability

Polynomial time Prediction Strategy with almost Optimal Mistake Probability Polynomial time Prediction Strategy with almost Optimal Mistake Probability Nader H. Bshouty Department of Computer Science Technion, 32000 Haifa, Israel bshouty@cs.technion.ac.il Abstract We give the

More information

Clique is hard on average for regular resolution

Clique is hard on average for regular resolution Clique is hard on average for regular resolution Ilario Bonacina, UPC Barcelona Tech July 27, 2018 Oxford Complexity Day Talk based on a joint work with: A. Atserias S. de Rezende J. Nordstro m M. Lauria

More information

Lower Bounds for Bounded Depth Frege Proofs via Buss-Pudlák Games

Lower Bounds for Bounded Depth Frege Proofs via Buss-Pudlák Games Lower Bounds for Bounded Depth Frege Proofs via Buss-Pudlák Games Eli Ben-Sasson Prahladh Harsha Abstract We present a simple proof of the bounded-depth Frege lower bounds of Pitassi et. al. and Krajíček

More information

Upper and Lower Bounds for Tree-like. Cutting Planes Proofs. Abstract. In this paper we study the complexity of Cutting Planes (CP) refutations, and

Upper and Lower Bounds for Tree-like. Cutting Planes Proofs. Abstract. In this paper we study the complexity of Cutting Planes (CP) refutations, and Upper and Lower Bounds for Tree-like Cutting Planes Proofs Russell Impagliazzo Toniann Pitassi y Alasdair Urquhart z UCSD UCSD and U. of Pittsburgh University of Toronto Abstract In this paper we study

More information

Degree and Sensitivity: tails of two distributions *

Degree and Sensitivity: tails of two distributions * Electronic Colloquium on Computational Complexity, Report No. 69 (2016) Degree and Sensitivity: tails of two distributions * Parikshit Gopalan Microsoft Research parik@microsoft.com Rocco A. Servedio Columbia

More information

Bias/variance tradeoff, Model assessment and selec+on

Bias/variance tradeoff, Model assessment and selec+on Applied induc+ve learning Bias/variance tradeoff, Model assessment and selec+on Pierre Geurts Department of Electrical Engineering and Computer Science University of Liège October 29, 2012 1 Supervised

More information

25 Minimum bandwidth: Approximation via volume respecting embeddings

25 Minimum bandwidth: Approximation via volume respecting embeddings 25 Minimum bandwidth: Approximation via volume respecting embeddings We continue the study of Volume respecting embeddings. In the last lecture, we motivated the use of volume respecting embeddings by

More information

An Introduction to Randomized algorithms

An Introduction to Randomized algorithms An Introduction to Randomized algorithms C.R. Subramanian The Institute of Mathematical Sciences, Chennai. Expository talk presented at the Research Promotion Workshop on Introduction to Geometric and

More information

Quantum boolean functions

Quantum boolean functions Quantum boolean functions Ashley Montanaro 1 and Tobias Osborne 2 1 Department of Computer Science 2 Department of Mathematics University of Bristol Royal Holloway, University of London Bristol, UK London,

More information