A Switching Lemma Tutorial. Benjamin Rossman University of Toronto and NII, Tokyo
|
|
- Anna Norris
- 5 years ago
- Views:
Transcription
1 A Switching Lemma Tutorial Benjamin Rossman University of Toronto and NII, Tokyo
2 homage to Paul Beame s excellent A Switching Lemma Primer A Switching Lemma Tutorial* Benjamin Rossman University of Toronto and NII, Tokyo
3 Overview 1. Basic Concepts models of computa?on (AC 0, DNFs/CNFs, decision trees) random restric?on R p 2. The Classic Switching Lemma (Hastad 86) 3. PARITY Lower Bound 4. Affine RestricGons Tsei?n expander switching lemma (joint with Pitassi, Servedio and Tan)
4 Models of Computa?on AC 0 circuits / formulas DNFs / CNFs (i.e. depth-2 formulas) Decision trees Canonical decision tree of a DNF / CNF
5 AC 0 Circuits size = # of gates depth = # layers of gates x 1 x 1 x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5
6 AC 0 Circuits size = # of gates depth = # layers of gates The complexity class AC 0 refers to (sequences of) Boolean func?ons {0,1} n {0,1} computable by poly(n)-size, constant-depth circuits. x 1 x 1 x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5
7 Depth-2 (DNFs and CNFs) DNF = disjunc?ve normal form (OR-AND formula) CNF = conjunc?ve normal form (AND-OR formula) width = max # of variables in a term/clause k
8 Depth-2 (DNFs and CNFs) k-dnf = width-k DNF k-cnf = width-k CNF k
9 Decision Trees x 3 depth 3 x 1 x 7 x 4 x 7 x 1 x
10 Decision Trees The decision-tree depth of a Boolean func?on f : {0,1} n {0,1} denoted DT depth (f), is the minimum depth of a decision tree that computes f DT depth (PARITY n ) = DT depth (AND n ) = n DT depth (f) = 0 f is constant
11 Decision Tree to DNF ( x 3 x 1 x 4 ) ( x 3 x 1 x 7 ) (x 3 x 7 x 1 ) (x 3 x 7 x 8 ) x 1 x 3 x 7 x 4 x 7 x 1 x
12 Decision Tree to DNF ( x 3 x 1 x 4 ) x 3 x 1 x 7 x 4 x 7 x 1 x
13 Decision Tree to DNF ( x 3 x 1 x 4 ) ( x 3 x 1 x 7 ) x 3 x 1 x 7 x 4 x 7 x 1 x
14 Decision Tree to DNF ( x 3 x 1 x 4 ) ( x 3 x 1 x 7 ) (x 3 x 7 x 1 ) x 3 x 1 x 7 x 4 x 7 x 1 x
15 Decision Tree to DNF ( x 3 x 1 x 4 ) ( x 3 x 1 x 7 ) (x 3 x 7 x 1 ) (x 3 x 7 x 8 ) x 1 x 3 x 7 x 4 x 7 x 1 x
16 (x 1 x 2 x 3 ) ( x 2 x 4 ) (x 1 x 3 x 7 ) Canonical DT of a DNF
17 (x 1 x 2 x 3 ) ( x 2 x 4 ) (x 1 x 3 x 7 ) Canonical DT of a DNF
18 Canonical DT of a DNF (x 1 x 2 x 3 ) x 1 x 2 x 3 1
19 Canonical DT of a DNF (x 1 x 2 x 3 ) ( x 2 x 4 ) x 1 x 2 x 4 1 x 4 1 x 2 x 2 x 4 1 x 3 x 2 1
20 Canonical DT of a DNF (x 1 x 2 x 3 ) ( x 2 x 4 ) x 1 x 2 x 4 1 x 4 1 x 2 x 2 x 4 1 x 3 x 2 1
21 Canonical DT of a DNF (x 1 x 2 x 3 ) ( x 2 x 4 ) x 1 x 2 x 4 1 x 4 x 2 1 x 3 1
22 Canonical DT of a DNF (x 1 x 2 x 3 ) ( x 2 x 4 ) x 1 (x 1 x 3 x 7 ) x 2 x x 3 x 7 x 4 x 2 1 x 7 1 x 3 1
23 Canonical DT of a DNF (x 1 x 2 x 3 ) ( x 2 x 4 ) x 1 (x 1 x 3 x 7 ) x 2 x x 3 x 7 0 x 4 0 x 2 1 x 7 1 x 3 0 1
24 Restric?ons a.k.a. func?ons [n] {0,1, } a.k.a. subcubes of {0,1} n
25 Restric?ons Consider a Boolean func?on f : {0,1} n {0,1} A restric3on (w.r.t. the variables of f) is a func?on R : [n] {0,1, }
26 Restric?ons Consider a Boolean func?on f : {0,1} n {0,1} A restric3on (w.r.t. the variables of f) is a func?on R : [n] {0,1, } [n] = {1,,n}, which we iden?fy with the variables x 1,,x n of f(x)
27 Restric?ons Consider a Boolean func?on f : {0,1} n {0,1} A restric3on (w.r.t. the variables of f) is a func?on R : [n] {0,1, } R(i) = 0 or 1 means that the variable x i is fixed to 0 or 1
28 Restric?ons Consider a Boolean func?on f : {0,1} n {0,1} A restric3on (w.r.t. the variables of f) is a func?on R : [n] {0,1, } R(i) = means that the variable x i is free (unrestricted)
29 Restric?ons f : {0,1} n {0,1} R : [n] {0,1, } Applying R to f, we get a Boolean func?on f R : {0,1} Stars(R) {0,1} R f R ( ) f ( )
30 Restric?ons Restric?ons are applied syntac3cally to DNFs / CNFs / decision trees / AC 0 circuits, etc.
31 Restric?ng an AC 0 Circuit Consider R = { x 1 1 } x 1 x 1 x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5
32 Restric?ng an AC 0 Circuit Consider R = { x 1 1 } i.e., R(1) = 1 and R(i) = for all i 1 x 1 x 1 x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5
33 Restric?ng an AC 0 Circuit Consider R = { x 1 1 } x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5
34 Restric?ng an AC 0 Circuit Consider R = { x 1 1 } x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5
35 Restric?ng an AC 0 Circuit Consider R = { x 1 1 } 1 0 x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5
36 Restric?ng an AC 0 Circuit Gate Elimina9on Method Consider R = { x 1 1 } N to 1+1/exp(d) lower bound against depth-d circuits via determinis3c restric3ons [Chaudhuri & Radhakrishnan 96] 1 0 x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5
37 Restric?ng a DNF R = { x 1 1, x 4 0 } x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 7 x 8 x 9 x 4 x 7 x 9
38 Restric?ng a DNF R = { x 1 1, x 4 0 } 1 x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 7 x 8 x 9 x 4 x 7 x 9
39 Restric?ng a DNF R = { x 1 1, x 4 0 } 1 x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 7 x 8 x 9 x 4 x 7 x 9
40 Restric?ng a DNF R = { x 1 1, x 4 0 } 1 x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 7 x 8 x 9 x 4 x 7 x 9
41 Restric?ng a DNF R = { x 1 1, x 4 0 } 1 x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 7 x 8 x 9 x 4 x 7 x 9 R x 2 x 3 x 2 x 5 x 7 x 8 x 9 x 7 x 9
42 Restric?ng a Decision Tree Again co R = { x 1 1 } x 3 x 1 x 7 x 4 x 7 x 1 x
43 Restric?ng a Decision Tree Again co R = { x 1 1 } x 3 x 1 x 7 x 4 x 7 x 1 x
44 Restric?ng a Decision Tree Again co R = { x 1 1 } x 3 R 0 x 3 1 x 1 x 7 x 7 x 7 x 4 x 7 x 1 x x 8 1 0
45 The Random Restric?on R p Throughout this talk, random objects in boldface
46 The Random Restric?on R p For 0 p 1, let R p be the random restric3on with R p (i) = with prob. p 0 with prob. (1-p)/2 1 with prob. (1-p)/2 independently for each variable index i [n]
47 Effect of R p R p simplifies Boolean func?ons represented by small decision trees / AC 0 circuits / demorgan formulas, [Subbotovskaya 61, ] Certain Boolean func?ons, like PARITY, maintain their complexity under R p => lower bounds!
48 Effect of R p on Boolean Func?ons R p kills AND n with very high probability (for p 1/2): Pr[ AND n R p 0 ] 1 (1/2) O(n)
49 Effect of R p on Boolean Func?ons R p kills AND n with very high probability (for p 1/2): Pr[ AND n R p 0 ] 1 (1/2) O(n) On the other hand, for any R with k stars, PARITY n R PARITY k or 1 PARITY k In par?cular, DT depth (PARITY n R p ) pn with high prob
50 Effect of R p on Decision Trees Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l
51 Effect of R p on Decision Trees Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l 2 l when p 1/4k
52 Effect of R p on DNFs Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l Hastad s Switching Lemma (1986) If F is a k-dnf (i.e. OR of depth-k decision trees), then Pr[ DT depth (F R p ) l ] (5pk) l
53 Effect of R p on DNFs Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t 2 l when R p ) p l ] 1/10k (2pk) l Hastad s Switching Lemma (1986) If F is a k-dnf (i.e. OR of depth-k decision trees), then Pr[ DT depth (F R p ) l ] (5pk) l
54 Effect of R p on DNFs Corollary (usual statement of the S.L.) If F is a k-dnf, then Pr[F R p is not equivalent to an l-cnf ] (5pk) l Hastad s Switching Lemma (1986) If F is a k-dnf (i.e. OR of depth-k decision trees), then Pr[ DT depth (F R p ) l ] (5pk) l
55 Effect of R p on DNFs CNFs Corollary (usual statement of the S.L.) If F is a k-cnf, then Pr[F R p is not equivalent to an l-dnf ] (5pk) l Hastad s Switching Lemma (1986) If F is a k-cnf (i.e. AND of depth-k decision trees), then Pr[ DT depth (F R p ) l ] (5pk) l
56 Effect of R p on DNFs k R p w.h.p. for small p l
57 Depth Reduc?on unbnd'd fan-in k k k k k k
58 Depth Reduc?on Apply the Switching Lemma to each gate (take a union bound over failure probability) k k k k k k
59 Depth Reduc?on Apply the Switching Lemma to each gate (take a union bound over failure probability) l l l l l l
60 Depth Reduc?on two layers of -gates l l l l l l
61 Depth Reduc?on l l l l l l
62 Depth Reduc?on Applying the S.L. to the boaom-2 layers k k k k k k l l l l l l
63 Depth Reduc?on Applying the S.L. to the boaom-2 layers k k k k k k typically, k = l n 1/d l l l l l l
64 Depth Reduc?on unbnd'd fan-in depth-k dec. trees
65 Hastad s Switching Lemma (decision tree version) If T 1,, T m are depth-k decision trees, then Pr[ DT depth ((T 1 T m ) R p ) l ] (5pk) l unbnd'd fan-in depth-k dec. trees
66 Hastad s Switching Lemma (decision tree version) If T 1,, T m are depth-k decision trees, then Pr[ DT depth ((T 1 T m ) R p ) l ] (5pk) l depth-l dec. trees
67 Depth Reduc?on applying the S.L. to the boaom layer of gates
68 PARITY Lower Bound
69 Hastad 86 Depth d+1 circuits for PARITY require size exp(ω(n 1/d ))
70 Hastad 86 Depth d+1 circuits for PARITY require size exp(ω(n 1/d )) (Trivial) Upper Bound PARITY has depth d+1 circuits of size exp(o(n 1/d )) d 2 n 1/d n 1 1/d n 1 1/d n 1 1/d
71 Hastad 86 Depth d+1 circuits for PARITY require size exp(ω(n 1/d )) (Trivial) Upper Bound PARITY has depth d+1 circuits of size exp(o(n 1/d )) d+1 n
72 PARITY Lower Bound x 1 x 1 x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5
73 PARITY Lower Bound depth-1 decision trees on n variables
74 PARITY Lower Bound R p -random restric?on ρ 1 : [n] {0,1, } depth-1 decision trees on n variables
75 PARITY Lower Bound Apply the Switching Lemma to each bo om-level gate Take a union bound over failure probabili?es depth-1 decision trees on n variables
76 PARITY Lower Bound Apply the Switching Lemma to each bo om-level gate Take a union bound over failure probabili?es low-depth decision trees on pn variables
77 PARITY Lower Bound 1 O(ε) frac?on have depth 0 O(ε) frac?on have depth 1 O(ε 2 ) frac?on have depth 2 O(ε l ) have depth l low-depth decision trees on pn variables
78 PARITY Lower Bound low-depth decision trees on pn variables
79 PARITY Lower Bound R p -random restric?on ρ 2 : Stars(ρ 1 ) {0,1, } low-depth decision trees on pn variables
80 PARITY Lower Bound low-depth decision trees on p 2 n variables
81 PARITY Lower Bound R p -random restric?on ρ 3 : Stars(ρ 2 ) {0,1, } low-depth decision trees on p 2 n variables
82 PARITY Lower Bound low-depth decision trees on p 3 n variables
83 PARITY Lower Bound R p -random restric?on ρ 4 : Stars(ρ 3 ) {0,1, } low-depth decision trees on p 3 n variables
84 PARITY Lower Bound depth-0 decision tree (i.e. constant funcgon) on p 4 n variables
85 PARITY Lower Bound depth-0 decision tree (i.e. constant funcgon) on p 4 n variables w.h.p., output is fixed to 0 or 1 a~er restric?ons ρ 1 ρ 2 ρ 3 ρ 4 (equivalent to R p 4) If p 4 n 1, this shows the original circuit does not compute PARITY n
86 PARITY Lower Bound Op?mal choice of parameters gives a?ght exp(ω(n 1/d )) lower bound for depth d+1 circuits x 1 x 1 x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5
87 Formulas vs. Circuits
88 AC 0 Circuits x 1 x 1 x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5
89 AC 0 Formulas x 3 x 7 x 4 x 1 x 3 x 2 x 11 x 1 x 2 x 5 x 4 x 2 x 1
90 Fact: Every depth d+1 circuit of size S is equivalent to a depth d+1 formula of size at most S d
91 PARITY Lower Bound Hastad 86 Depth d+1 circuits for PARITY require size exp(ω(n 1/d )) R. 15 Depth d+1 formulas for PARITY have size exp(ω(dn 1/d ))
92 PARITY Lower Bound Hastad 86 Depth d+1 circuits for PARITY require size exp(ω(n 1/d )) Poly-size circuits for PARITY have depth log n log log n + O(1) R. 15 Depth d+1 formulas for PARITY have size exp(ω(dn 1/d )) Poly-size formulas for PARITY have depth Ω(log n)
93 Two Views of R p : [n] {0,1, } (1) Independently, for each coordinate i [n], set R p (i) = with prob. p 0 with prob. (1-p)/2 1 with prob. (1-p)/2 (2) σ {0,1} n uniform random assignment iτ [0,1] n uniform random?me-stamp R p (i) = if τ i p σ i if τ i > p
94 1 τ 0 σ σ {0,1} n uniform random assignment iτ [0,1] n uniform random?me-stamp R p (i) = if τ i p σ i if τ i > p
95 1 τ?me p runs from 1 to 0 0 σ σ {0,1} n uniform random assignment iτ [0,1] n uniform random?me-stamp R p (i) = if τ i p σ i if τ i > p
96 1 start?me p = 1 τ 0 σ R p σ {0,1} n uniform random assignment iτ [0,1] n uniform random?me-stamp R p (i) = if τ i p σ i if τ i > p
97 1 start?me p = 1 τ 0 σ R p σ {0,1} n uniform random assignment iτ [0,1] n uniform random?me-stamp R p (i) = if τ i p σ i if τ i > p
98 1 p = 0.99 τ 0 σ R p 0 σ {0,1} n uniform random assignment iτ [0,1] n uniform random?me-stamp R p (i) = if τ i p σ i if τ i > p
99 1 p = 0.5 τ 0 σ R p σ {0,1} n uniform random assignment iτ [0,1] n uniform random?me-stamp R p (i) = if τ i p σ i if τ i > p
100 1 p = 0.1 τ 0 σ R p σ {0,1} n uniform random assignment iτ [0,1] n uniform random?me-stamp R p (i) = if τ i p σ i if τ i > p
101 1 end?me p = 0 τ 0 σ R p σ {0,1} n uniform random assignment iτ [0,1] n uniform random?me-stamp R p (i) = if τ i p σ i if τ i > p
102 1 0 x 1 x 1 x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5
103 1 p 0 R p x 1 x 1 x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5
104 1 p 0 p 2 R p 2 R p x 1 x 1 x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5
105 1 p 0 p 2 p 3 R p 3 R p 2 R p x 1 x 1 x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5
106 1 p 0 p 2 p 3 p 4 R p 4 R p 3 R p 2 R p x 1 x 1 x 2 x 2 x 3 x 3 x 4 x 4 x 5 x 5
107 1 0
108 1 0 Each sub-formula F will have its own stopping?me q(f) q(f) is a random variable in [0,1], which is determined by σ and τ
109 1 q(g) 0 G
110 1 q(h) 0 H
111 1 q(i) 0 I
112 1 q(j) 0 J
113 1 0 q(f) F
114 1 p 0
115 1 p 0 R p
116 1 p 0 decision trees
117 1 p 0 k
118 1 p 0 F k
119 1 p 0 q(f) F q(f) = p / 14k k
120 Hastad s Switching Lemma If F is k-dnf or k-cnf, then Pr[ DT depth (f R 1/14k ) l ] e l Stopping Time Version For every Boolean formula F, Pr[ DT depth (F R q(f) ) l ] e l
121 Technical Main Lemma (tail bound for q(f)) If F has depth d+1, then for all 0 λ 1, C Pr[ q(f) λ ] size(f) d exp(ω(dλ 1/d )) The constant C < 8 is defined by 1+ 1X i=0 1 1 exp(e i 1 (i + 1)e 2 ) + X i=0 1X j=0 1 exp((j + 1)e i 1 (i + j + 2)e 2 )
122 Technical Main Lemma (tail bound for q(f)) If F has depth d+1, then for all 0 λ 1, C Pr[ q(f) λ ] size(f) d exp(ω(dλ 1/d )) Corollary Depth d+1 formulas for PARITY have size exp(ω(dn 1/d ) O(d))
123 Proof of Hastad s Switching Lemma
124 Fix a k-dnf F and l 1 Switching Lemma: Pr[ depth(candt(f R p )) l ] = O(pk) l
125 Fix a k-dnf F and l 1 BAD := { restric?ons R depth(candt(f R)) l } Switching Lemma: Pr[ depth(candt(f R p )) l ] = O(pk) l
126 Fix a k-dnf F and l 1 BAD := { restric?ons R depth(candt(f R)) l } Goal. Pr[ R p BAD ] = O(pk) l Switching Lemma: Pr[ depth(candt(f R p )) l ] = O(pk) l
127 Fix a k-dnf F and l 1 BAD := { restric?ons R depth(candt(f R)) l } Goal. Pr[ R p BAD ] = O(pk) l Key idea. We associate each R BAD with a restric?on R* such that
128 Fix a k-dnf F and l 1 BAD := { restric?ons R depth(candt(f R)) l } Goal. Pr[ R p BAD ] = O(pk) l Key idea. We associate each R BAD with a restric?on R* such that 1 Stars(R*) = Stars(R) l R* R* (l = 3)
129 Fix a k-dnf F and l 1 BAD := { restric?ons R depth(candt(f R)) l } Goal. Pr[ R p BAD ] = O(pk) l Key idea. We associate each R BAD with a restric?on R* such that 1 Stars(R*) = Stars(R) l in par?cular, Pr[ R p = R* ] / Pr[ R p = R ] = ((1-p)/2p) l
130 Fix a k-dnf F and l 1 BAD := { restric?ons R depth(candt(f R)) l } Goal. Pr[ R p BAD ] = O(pk) l Key idea. We associate each R BAD with a restric?on R* such that 1 Stars(R*) = Stars(R) l in par?cular, Pr[ R p = R* ] / Pr[ R p = R ] = ((1-p)/2p) l for any restric?on ρ with s stars, Pr[ R p = ρ ] = p s (1 p) n s
131 Fix a k-dnf F and l 1 BAD := { restric?ons R depth(candt(f R)) l } Goal. Pr[ R p BAD ] = O(pk) l Key idea. We associate each R BAD with a restric?on R* such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1
132 Fix a k-dnf F and l 1 BAD := { restric?ons R depth(candt(f R)) l } Goal. Pr[ R p BAD ] = O(pk) l Key idea. We associate each R BAD with a restric?on R* such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1 any restric?on ρ equals R* for at most (4k) l dis?nct R BAD
133 Key idea. We associate each R BAD with a restric?on R* such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1
134 Key idea. We associate each R BAD with a restric?on R* such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1 Pr[ R p BAD ] = R BAD Pr[ R p = R ]
135 Key idea. We associate each R BAD with a restric?on R* such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1 Pr[ R p BAD ] = R BAD Pr[ R p = R ] = R BAD (2p/(1-p)) l Pr[ R p = R* ] Pr[ R p = R ] = (2p/(1-p)) l Pr[ R p = R* ]
136 Key idea. We associate each R BAD with a restric?on R* such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1 Pr[ R p BAD ] = R BAD Pr[ R p = R ] = R BAD (2p/(1-p)) l Pr[ R p = R* ] (4p) l R BAD Pr[ R p = R* ] w.l.o.g. p 1/2
137 Key idea. We associate each R BAD with a restric?on R* such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1 Pr[ R p BAD ] = R BAD Pr[ R p = R ] = R BAD (2p/(1-p)) l Pr[ R p = R* ] (4p) l R BAD Pr[ R p = R* ] (4p) l (4k) l Pr[ R p {R* R BAD} ]
138 Key idea. We associate each R BAD with a restric?on R* such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1 Pr[ R p BAD ] = R BAD Pr[ R p = R ] = R BAD (2p/(1-p)) l Pr[ R p = R* ] (4p) l R BAD Pr[ R p = R* ] (4p) l (4k) l Pr[ R p {R* R BAD} ] Pr[ ] 1
139 Key idea. We associate each R BAD with a restric?on R* such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1 Pr[ R p BAD ] = R BAD Pr[ R p = R ] = R BAD (2p/(1-p)) l Pr[ R p = R* ] (4p) l R BAD Pr[ R p = R* ] (4p) l (4k) l Pr[ R p {R* R BAD} ] (16pk) l Q.E.D.
140 Switching Lemma: Pr[ depth(candt(f R p )) l ] = O(pk) l Pr[ R p BAD ] = R BAD Pr[ R p = R ] = R BAD (2p/(1-p)) l Pr[ R p = R* ] (4p) l R BAD Pr[ R p = R* ] (4p) l (4k) l Pr[ R p {R* R BAD} ] (16pk) l Q.E.D.
141 Switching Lemma: Pr[ depth(candt(f R p )) l ] = O(pk) l Pr[ R p BAD ] = R BAD Pr[ R p = R ] = R BAD (2p/(1-p)) l Pr[ R p = R* ] more (4p) careful l R BAD analysis Pr[ R p = gives R* ] (5pk) l (w.l.o.g. p (4p) l (4k) l Pr[ R p {R* R BAD} ] (16pk) l Q.E.D.
142 Key idea. We associate each R BAD with a restric?on R* such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1
143 Key idea. We associate each R BAD with a pair (R*,Code(R)) such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1 3 4
144 Key idea. We associate each R BAD with a pair (R*,Code(R)) such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1 3 Code(R) ({0,1} 2 [k]) l 4
145 Key idea. We associate each R BAD with a pair (R*,Code(R)) such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1 3 Code(R) ({0,1} 2 [k]) l 4 in par?cular, Code(R) has (4k) l possible values
146 Key idea. We associate each R BAD with a pair (R*,Code(R)) such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1 3 Code(R) ({0,1} 2 [k]) l 4 the map R (R*,Code(R)) is 1-to-1
147 Key idea. We associate each R BAD with a pair (R*,Code(R)) such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1 3 Code(R) ({0,1} 2 [k]) l 4 the map R (R*,Code(R)) is 1-to-1 3 & 4 imply 2
148 Key idea. We associate each R BAD with a pair (R*,Code(R)) such that 1 Stars(R*) = Stars(R) l 2 the map R R* is (4k) l -to-1 3 Code(R) ({0,1} 2 [k]) l 4 the map R (R*,Code(R)) is 1-to-1 intui?vely, Code(R) is a recipe for inver?ng R (R*,Code(R)) given knowledge of R*
149 R (R*,Code(R))
150 k = 3, l = 4 R (R*,Code(R)) F = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7
151 k = 3, l = 4 R (R*,Code(R)) F = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 }
152 k = 3, l = 4 R (R*,Code(R)) 1 1 F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 }
153 k = 3, l = 4 R (R*,Code(R)) 1 1 F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 }
154 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } x 2 CanDT(F R) = x 7 x 3 1 x 5 x
155 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R BAD, since CanDT(F R) l 4 x 2 x 7 x 3 1 x 5 x
156 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } x 2 x 7 x 3 1 x 5 x
157 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 } x 2 x 7 x 3 1 x 5 x
158 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 } x 2 x 7 x 3 1 x 5 x
159 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 } x 2 x 7 x 3 1 x 5 x
160 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 } x 2 x 7 x 3 1 x 5 x
161 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 } x 2 x 7 x 3 1 x 5 x
162 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 } Code(R) says: find the first sa?sfied term of F R* the long path begins with x 2 1, x 3 1 (i.e. var 2 of term 1 and var 3 of term 1)
163 k = 3, l = 4 R (R*,Code(R)) 1 F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } Fix variables according to the beginning of the long path R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 }
164 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 } x 2 x 7 x 3 1 x 5 x
165 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 } x 2 x 7 x 3 1 x 5 x
166 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 } x 2 x 7 x 3 1 x 5 x
167 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 } x 2 x 7 x 3 1 x 5 x
168 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 } Code(R) next says: find the next sa?sfied term of F R*(overwri?ng x 2 1, x 3 1) the long path con?nues x 5 0 (i.e. var 3 of term 0)
169 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } 0 R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } Fix variables according to the long path R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 }
170 k = 3, l = 4 R (R*,Code(R)) F R = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 } x 2 x 7 x 3 1 x 5 x
171 k = 3, l = 4 R (R*,Code(R)) F = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } Code(R) ({0,1} 2 [k]) l given knowledge of R* (and F), follow these instruc?ons to recover R (and along the way R~) R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 }
172 k = 3, l = 4 R (R*,Code(R)) F = x 1 x 2 x 3 x 1 x 3 x 5 x 2 x 4 x 5 x 3 x 4 x 6 x 1 x 4 x 7 R = { x 1 1, x 4 0 } R~ = { x 1 1, x 4 0, x 2 1, x 3 1, x 5 0, x 7 1 } R* = { x 1 1, x 4 0, x 2 1, x 3 0, x 5 1, x 7 0 } Code(R) ({0,1} 2 [k]) l given knowledge of R* (and F), follow these instruc?ons to recover R (and along the way R~) ü R* has l fewer stars than R ü R (R*,Code(R)) is 1-to-1
173 A Different Approach
174 A Different Approach Hastad s Switching Lemma If F is a k-dnf, then Pr[ depth(candt(f R p )) l ] = O(pk) l Weak (but flexible) Switching Lemma If F is a k-dnf, then Pr[ depth(candt(f) R p ) l ] = O(pk2 k ) l
175 Shows that PARITY has depth-d circuit size log n d 2 exp(ω(( ) )) Weak (but flexible) Switching Lemma If F is a k-dnf, then Pr[ depth(candt(f) R p ) l ] = O(pk2 k ) l
176 Proof generalizes to affine restric3ons! Weak (but flexible) Switching Lemma If F is a k-dnf, then Pr[ depth(candt(f) R p ) l ] = O(pk2 k ) l
177 A Different Approach 1. Decision Tree Switching Lemma 2. k-clipped Decision Trees 3. Arbitrary Distribu?on of Stars 4. Switching Lemma for Affine Restric3ons 5. Tsei?n Expander Switching Lemma
178 A Different Approach 1. Decision Tree Switching Lemma 2. k-clipped Decision Trees 3. Arbitrary Distribu?on of Stars 4. Switching Lemma for Affine Restric3ons 5. Tsei?n Expander Switching Lemma
179 Binomial Distribu?on For a finite set N, let S p N Denote S is a p-biased random subset of N, i.e., Pr[ S = S ] = p S (1 p) N S For a natural number n and 0 p 1, the binomial random variable Bin(n,p) is equivalent to S where S p [n]
180 Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l
181 Consider a random branch of T: x 3 x 1 x 2 1/4 x 4 x 7 x 7 1/8 x 2 1/16 1/16 1/8 1/8 1/8 1/8
182 Consider a random branch of T: x 3 x 1 x 2 1/4 x 4 x 7 x 7 1/8 x 2 1/16 1/16 1/8 1/8 1/8 1/8
183 Let β(t) [n] be the set of variables queried on a random branch of T x 3 x 1 x 2 1/4 x 4 x 7 x 7 1/8 x 2 1/16 1/16 1/8 1/8 1/8 1/8
184 #β(t) = the number of variables queried on a random branch of T x 3 x 1 x 2 1/4 x 4 x 7 x 7 1/8 x 2 1/16 1/16 1/8 1/8 1/8 1/8
185 #β(t) = the number of variables queried on a random branch of T x 3 Obs: If depth(t) k, then Pr[ #β(t) k ] 2 k x 1 x 2 1/4 x 4 x 7 x 7 1/8 x 2 1/16 1/16 1/8 1/8 1/8 1/8
186 Claim. The following random variables have the same distribu?on: 1. β(t R p ) the set of variables queried on a random branch of T R p 2. S p β(t) a p-biased subset of the vars. on a random branch of T
187 Claim. The following random variables have the same distribu?on: 1. β(t R p ) #β(t R p ) the set of variables queried on a random branch of T R p 2. S p β(t) Bin(#β(T), p) a p-biased subset of the vars. on a random branch of T
188 1. β(t R p ) 2. S p β(t) T
189 1. β(t R p ) 2. S p β(t) Stars(R p )
190 1. β(t R p ) 2. S p β(t) R p
191 1. β(t R p ) 2. S p β(t) T R p
192 1. β(t R p ) 2. S p β(t) T R p
193 1. β(t R p ) 2. S p β(t) T R p
194 1. β(t R p ) 2. S p β(t) T R p
195 1. β(t R p ) 2. S p β(t) β(t R p )
196 1. β(t R p ) 2. S p β(t) β(t R p )
197 1. β(t R p ) 2. S p β(t) β(t R p )
198 1. β(t R p ) 2. S p β(t) β(t R p )
199 1. β(t R p ) 2. S p β(t) β(t R p )
200 1. β(t R p ) 2. S p β(t) T
201 1. β(t R p ) 2. S p β(t) β(t)
202 1. β(t R p ) 2. S p β(t) β(t)
203 1. β(t R p ) 2. S p β(t) S p β(t)
204 1. β(t R p ) 2. S p β(t) 1. #β(t R p ) 2. Bin(β(T), p)
205 Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l
206 Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l Pr[ depth(t R p ) l ] recall: depth(t ) l Pr[ #β(t ) l ] 2 l
207 Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l Pr[ depth(t R p ) l ] Pr R p [ Pr β(t R p) [ #β(t R p ) l ] 2 l ] p p recall: depth(t ) l Pr[ #β(t ) l ] 2 l
208 Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l Pr[ depth(t R p ) l ] Pr R p [ Pr β(t R p) [ #β(t R p ) l ] 2 l ] p p
209 Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l Pr[ depth(t R p ) l ] Pr R p [ Pr β(t R p) [ #β(t R p ) l ] 2 l ] p 2 l Ex R p [ Pr β(t R p) [ #β(t R p ) l ] ] p p p Markov s inequality
210 Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l Pr[ depth(t R p ) l ] Pr R p [ Pr β(t R p) [ #β(t R p ) l ] 2 l ] p p 2 l Ex R p [ Pr β(t R p) [ #β(t R p ) l ] ] p p
211 Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l Pr[ depth(t R p ) l ] Pr R p [ Pr β(t R p) [ #β(t R p ) l ] 2 l ] p 2 l Pr[ #β(t R p ) l ] p
212 Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l Pr[ depth(t R p ) l ] Pr R p [ Pr β(t R p) [ #β(t R p ) l ] 2 l ] p 2 l Pr[ #β(t R p ) l ] = 2 l Pr[ Bin(#β(T), p) l ] p #β(t R p ) and Bin(#β(T), p) have the same distribu?on
213 Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l Pr[ depth(t R p ) l ] Pr R p [ Pr β(t R p) [ #β(t R p ) l ] 2 l ] p 2 l Pr[ #β(t R p ) l ] = 2 l Pr[ Bin(#β(T), p) l ] 2 l Pr[ Bin(k, p) l ] p #β(t) depth(t) k
214 Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l Pr[ depth(t R p ) l ] Pr R p [ Pr β(t R p) [ #β(t R p ) l ] 2 l ] p 2 l Pr[ #β(t R p ) l ] = 2 l Pr[ Bin(#β(T), p) l ] 2 l Pr[ Bin(k, p) l ] 2 l (pk) l p union bound
215 Decision Tree Switching Lemma If T is a depth-k decision tree, then Pr[ depth(t R p ) l ] (2pk) l Pr[ depth(t R p ) l ] Pr R p [ Pr β(t R p) [ #β(t R p ) l ] 2 l ] p 2 l Pr[ #β(t R p ) l ] = 2 l Pr[ Bin(#β(T), p) l ] 2 l Pr[ Bin(k, p) l ] 2 l (pk) l p Q.E.D.
216 A Different Approach 1. Decision Tree Switching Lemma 2. k-clipped Decision Trees 3. Arbitrary Distribu?on of Stars 4. Switching Lemma for Affine Restric3ons 5. Tsei?n Expander Switching Lemma
217 Recall: Canonical DT of a k-dnf (x 1 x 2 x 3 ) ( x 2 x 4 ) (x 1 x 3 x 7 ) x 4 1 x 2 1 x 1 x 3 x 7 x 4 x 2 1 x 7 1 x 3 1
218 Recall: Canonical DT of a k-dnf (x 1 x 2 x 3 ) ( x 2 x 4 ) (x 1 x 3 x 7 ) x 4 1 x 2 1 x 1 x 3 Obs: Every node has distance k from a leaf! x 7 x 4 x 2 1 x 7 1 x 3 1
219 Defini?on: A decision tree is k-clipped if every node has distance k from a leaf
220 Defini?on: A decision tree is k-clipped if every node has distance k from a leaf Lemma: If F is a k-dnf (or k-cnf), then CanDT(F) is k-clipped
221 Defini?on: A decision tree is k-clipped if every node has distance k from a leaf Lemma: If F is a k-dnf (or k-cnf), then CanDT(F) is k-clipped Lemma: If T is k-clipped, then Ex #β(t) l ( ) O(k2 k ) l
222 Defini?on: A decision tree is k-clipped if every node has distance k from a leaf #β(t) stochas?cally dominated by Lemma: If F is a k-dnf (or k-cnf), then the number of unbiased coin flips un?l CanDT(F) seeing is k consecu?ve k-clipped heads Lemma: If T is k-clipped, then Ex #β(t) l ( ) O(k2 k ) l
223 k-clipped Decision Tree Switching Lemma If T is a k-clipped decision tree, then Pr[ depth(t R p ) l ] = O(pk2 k ) l
224 k-clipped Decision Tree Switching Lemma If T is a k-clipped decision tree, then Pr[ depth(t R p ) l ] = O(pk2 k ) l Pr[ depth(t R p ) l ] Pr R p[ Pr β(t R p) [ #β(t R p ) l ] 2 l ] p 2 l Pr[ #β(t R p ) l ] p = 2 l Pr [ Bin(#β(T), p) l ] same as before
225 k-clipped Decision Tree Switching Lemma If T is a k-clipped decision tree, then Pr[ depth(t R p ) l ] = O(pk2 k ) l Pr[ depth(t R p ) l ] Pr R p[ Pr β(t R p) [ #β(t R p ) l ] 2 l ] p 2 l Pr[ #β(t R p ) l ] = 2 l Pr [ Bin(#β(T), p) l ] 2 l Ex[ p l ] p #β(t) l ( ) union bound
226 k-clipped Decision Tree Switching Lemma If T is a k-clipped decision tree, then Pr[ depth(t R p ) l ] = O(pk2 k ) l Lemma: If T is k-clipped, then Pr[ depth(t R p ) l ] Ex (#β(t)) Pr R p[ Pr β(t R p) [ #β(t O(k2 k R p ) ) l l l ] 2 l ] p 2 l Pr[ #β(t R p ) l ] = 2 l Pr [ Bin(#β(T), p) l ] 2 l Ex[ p l ] 2 l p l O(k2 k ) l p #β(t) l ( )
227 k-clipped Decision Tree Switching Lemma If T is a k-clipped decision tree, then Pr[ depth(t R p ) l ] = O(pk2 k ) l Pr[ depth(t R p ) l ] Pr R p[ Pr β(t R p) [ #β(t R p ) l ] 2 l ] p 2 l Pr[ #β(t R p ) l ] = 2 l Pr [ Bin(#β(T), p) l ] 2 l Ex[ p l ] 2 l p l O(k2 k ) l p #β(t) l ( ) Q.E.D.
228 A Different Approach 1. Decision Tree Switching Lemma 2. k-clipped Decision Trees 3. Arbitrary DistribuGon of Stars 4. Switching Lemma for Affine Restric3ons 5. Tsei?n Expander Switching Lemma
229 Let S [n] be an arbitrary distribu?on of stars
230 Let S [n] be an arbitrary distribu?on of stars Let u unif {0,1} n
231 Let S [n] be an arbitrary distribu?on of stars Let u unif {0,1} n Define random restric?on ρ S,u : [n] {0,1, } by ρ S,u (i) = S,u (i) = if i S, u i otherwise.
232 Let S [n] be an arbitrary distribu?on of stars Let u unif {0,1} n Define random restric?on ρ S,u : [n] {0,1, } by ρ S,u (i) = if i S, u i otherwise. If S p [n], then ρ S,u has the S,u has the same distribu?on as R p
233 Let S [n] be an arbitrary distribu?on of stars Defini?on: For p [0,1], we say that S is p-bounded if Pr[ J S ] p J for every J [n]
234 The following are equivalent: β(t) S 1. β(t ρ S,u ) 2. β(t) S
235 The following are equivalent: β(t) S 1. β(t ρ S,u ) 2. β(t) S in fact, this equivalence valid for fixed S Supp(S)
236 The following are equivalent: 1. β(t ρ S,u ) 2. β(t) S β(t) S If S is p-bounded, then Pr[ β(t) S l ] #β(t) l ( ) Ex[ p l ]
237 The following are equivalent: 1. β(t ρ S,u ) 2. β(t) S β(t) S Pr[ J S ] p l for every J [n] of size J = l If S is p-bounded, then Pr[ β(t) S l ] #β(t) l ( ) Ex[ p l ]
238 Theorem (Switching Lemma for ρ S,u ) If S is p-bounded and T is a k-clipped DT, then Pr[ depth(t ρ S,u ) l ] = O(pk2k k ) l
239 Theorem (Switching Lemma for ρ S,u ) If S is p-bounded and T is a k-clipped DT, then Pr[ depth(t ρ S,u ) l ] = O(pk2k k ) l Pr[ depth(t ρ S,u ) l ] Pr S,u [ Pr V(T ρ ) [ #β(t ρ S,u ) l ] 2 2 l l ] S,u 2 l Pr[ #β(t ρ S,u ) l ] = 2 l Pr [ β(t) S l ] #β(t) l ( ) 2 l Ex[ p l ] 2 l p l O(k2 k ) l Q.E.D.
240 A Different Approach 1. Decision Tree Switching Lemma 2. k-clipped Decision Trees 3. Arbitrary Distribu?on of Stars 4. Switching Lemma for Affine Restric9ons 5. Tsei?n Expander Switching Lemma
241 Restric?ons [n] {0,1, } are in 1-1 correspondence with subcubes of {0,1} n What about the complexity of F A (e.g. for F a k-dnf) for a random affine subspace of A {0,1} n?
242 Restric?ons [n] {0,1, } are in 1-1 correspondence with subcubes of {0,1} n What about the complexity of F A (e.g. for F a k-dnf) for a random affine subspace of A {0,1} n? A = V + u where V is a linear subspace of {0,1} n
243 Restric?ons [n] {0,1, } are in 1-1 correspondence with subcubes of {0,1} n What about the complexity of F A (e.g. for F a k-dnf) for a random affine subspace of A {0,1} n? Affine Switching Lemma If F is a k-dnf and A is a p-bounded random affine subset of {0,1} n, then Pr[ DT depth (F A) l ] = O(pk2 k ) l
244 Applying an Affine Restric?on Consider an affine set A {0,1} n For f : {0,1} n {0,1}, we define: f A : A {0,1}, DT depth (f A) = min. depth of a decision tree T such that f(x) = T(x) for all x A We also have a syntac3c opera?on T T A on decision trees
245 Applying an Affine Restric?on Example: A = { x {0,1} n x 1 x 3 } x 3 A 0 x 3 1 x 1 x 7 x 7 x 7 x 4 x 7 x 1 x x 8 1 0
246 Let S [n] be an arbitrary dist. of stars Let u unif {0,1} n Consider random restric?on ρ S,u (i) = S,u (i) = if i S, u i otherwise.
247 Let S [n] be an arbitrary dist. of stars Let V be an arbitrary dist. on linear subspaces of {0,1} n Let u unif {0,1} n Consider random restric?on ρ S,u (i) = S,u (i) = if i S, u i otherwise.
248 Let S [n] be an arbitrary dist. of stars Let V be an arbitrary dist. on linear subspaces of {0,1} n Let u unif {0,1} n Consider random restric?on ρ S,u (i) = Consider random affine space V + u if i S, u i otherwise.
249 Let S [n] be an arbitrary dist. of stars Let V be an arbitrary dist. on linear subspaces of {0,1} n Let u unif {0,1} n Consider random restric?on ρ S,u (i) = Consider random affine space V + u if i S, u i otherwise. V is the arrangement of stars
250 Let S [n] be an arbitrary dist. of stars Let V be an arbitrary dist. on linear subspaces of {0,1} n Let u unif {0,1} n Consider random restric?on ρ S,u (i) = Consider random affine space V + u if i S, u i otherwise. u gives the assignment to non-stars
251 Let S [n] be an arbitrary dist. of stars Let V be an arbitrary dist. on linear subspaces of {0,1} n Let u unif {0,1} n Consider random restric?on ρ S,u (i) = Consider random affine space V + u if i S, u i otherwise. {b 1,, b n } = standard basis for {0,1} n if V = Span{ b i i S }, then V + u is the subcube defined by ρ S,u
252 Defini?on A random set S [n] is p-bounded if, for all J [n], Pr[ J S ] p J Defini?on A random linear subspace V {0,1} n is p-bounded if, for all J [n], Pr[ V sha ers J ] p J
253 Defini?on A random set S [n] is p-bounded if, for all J [n], Pr[ J S ] p J Defini?on A random linear subspace V {0,1} n is p-bounded if, for all J [n], Pr[ V sha ers J ] p J I J v V such that v i = 1 for all i I v j = 0 for all j J I
254 Defini?on V sha ers J [n] A random set S [n] is p-bounded if, for all J [n], J is independent Pr[ J S ] p in J the matroid associated with V Defini?on A random linear subspace V {0,1} n is p-bounded if, for all J [n], Pr[ V sha ers J ] p J I J v V such that v i = 1 for all i I v j = 0 for all j J I
255 Defini?on A random set S [n] is p-bounded if, for all J [n], Pr[ J S ] p J Defini?on A random linear subspace V {0,1} n is p-bounded if, for all J [n], Pr[ V sha ers J ] p J equivalent if V = Span{ b i i S }
256 Suppose A = V + u where V is a p-bounded random linear subspace of {0,1} n u unif {0,1} n Theorem For every k-clipped decision tree T, Pr[ depth(t A) l ] = O(pk2 k ) l essen?ally the same proof as the Switching Lemma for ρ S,u
257 Suppose A = V + u where V is a p-bounded random linear subspace of {0,1} n u unif {0,1} n Theorem For every k-clipped decision tree T, Pr[ depth(t A) l ] = O(pk2 k ) l Corollary (Affine Switching Lemma) For every k-dnf F, Pr[ DT depth (F A) l ] = O(pk2 k ) l
258 Suppose A = V + u where V is a p-bounded random linear subspace of {0,1} n u unif {0,1} n Theorem For every k-clipped decision tree T, Pr[ depth(t A) l ] = O(pk2 k ) l Corollary (Affine Switching Lemma) For every k-dnf F, Pr[ DT depth (F A) l ] = O(pk2 k ) l Conjecture: O(pk) l
259 For every fixed V (and random u {0,1} n ), the following are equivalent: 1. β(t (V + u)) #β(t (V + u)) 2. greedy basis of β(t) in the V-matroid rank V (β(t))
260 Suppose A = V + u where V is a p-bounded random linear subspace of {0,1} n u unif {0,1} n Final generaliza?on: Replace the ambient space {0,1} n with an arbitrary affine space B
261 Suppose A = V + u where V is a p-bounded random linear subspace of B u unif B Final generaliza?on: Replace the ambient space {0,1} n with an arbitrary affine space B
262 Suppose A = V + u where V is a p-bounded random linear subspace of B u unif B A is a random affine subspace of B
263 Suppose A = V + u where V is a p-bounded random linear subspace of B u unif B Theorem For every B-independent k-clipped decision tree T, Pr[ depth(t A) l ] = O(pk2 k ) l
264 Suppose A = V + u where V is a p-bounded random linear subspace of B u unif B Theorem For every B-independent k-clipped decision tree T, Pr[ depth(t A) l ] = O(pk2 k ) l Every branch β [n] of T is sha ered by B (i.e. β is independent in the matroid associated with Lin(B) = { x y x,y B})
265 A Different Approach 1. Decision Tree Switching Lemma 2. k-clipped Decision Trees 3. Arbitrary Distribu?on of Stars 4. Switching Lemma for Affine Restric3ons 5. TseiGn Expander Switching Lemma
266 Tsei?n Expander Switching Lemma Theorem (Pitassi-R.-Servedio-Tan 16) Depth-d AC 0 Frege proofs of Tsei?n(3-regular log n d expander) require size exp(ω(( ) 2 )) Theorem (Beame-Pitassi-Impagliazzo 93, Ben-Sasson 02) Depth-d AC 0 Frege proofs of Tsei?n(3-regular expander) require size exp(ω(n 1/2 d ))
267 Tsei?n Expander Switching Lemma Theorem (Pitassi-R.-Servedio-Tan 16) Depth-d AC 0 Frege proofs of Tsei?n(3-regular log n d expander) require size exp(ω(( ) 2 )) Theorem (Beame-Pitassi-Impagliazzo 93, Ben-Sasson 02) Depth-d Improving AC 0 Frege the proofs AFFINE of SWITCHING Tsei?n(3-regular LEMMA would yield exp(ω(n 1/d expander) require size exp(ω(n 1/2 d )) ))
268 Suppose A = V + u where V is a p-bounded random linear subspace of {0,1} n u unif {0,1} n Affine Switching Lemma for k-dnfs For every k-dnf F, Pr[ DT depth (F A) l ] = O(pk2 k ) l Conjecture: O(pk) l
269 Stay Tuned Toni Pitassi: Expander S.L. & AC 0 -Frege lower bound Johan Hastad, Avishay Tal: Correla?on bounds from improved switching lemmas Johan Hastad, Rocco Servedio: Random projec?ons, depth hierarchy theorem Srikanth Srinivasan: Adap?ve random restric?ons (against AC 0 with few threshold gates)
270 Thanks!
CSC 2429 Approaches to the P vs. NP Question and Related Complexity Questions Lecture 2: Switching Lemma, AC 0 Circuit Lower Bounds
CSC 2429 Approaches to the P vs. NP Question and Related Complexity Questions Lecture 2: Switching Lemma, AC 0 Circuit Lower Bounds Lecturer: Toniann Pitassi Scribe: Robert Robere Winter 2014 1 Switching
More informationCS6840: Advanced Complexity Theory Mar 29, Lecturer: Jayalal Sarma M.N. Scribe: Dinesh K.
CS684: Advanced Complexity Theory Mar 29, 22 Lecture 46 : Size lower bounds for AC circuits computing Parity Lecturer: Jayalal Sarma M.N. Scribe: Dinesh K. Theme: Circuit Complexity Lecture Plan: Proof
More informationCS 161: Design and Analysis of Algorithms
CS 161: Design and Analysis of Algorithms NP- Complete I P, NP Polynomial >me reduc>ons NP- Hard, NP- Complete Sat/ 3- Sat Decision Problem Suppose there is a func>on A that outputs True or False A decision
More informationLecture 2: Proof of Switching Lemma
Lecture 2: oof of Switching Lemma Yuan Li June 25, 2014 1 Decision Tree Instead of bounding the probability that f can be written as some s-dnf, we estimate the probability that f can be computed by a
More informationCSCI 1010 Models of Computa3on. Lecture 02 Func3ons and Circuits
CSCI 1010 Models of Computa3on Lecture 02 Func3ons and Circuits Overview Func3ons and languages Designing circuits from func3ons Minterms and the DNF Maxterms and CNF Circuit complexity Algebra of Boolean
More informationLecture 4: LMN Learning (Part 2)
CS 294-114 Fine-Grained Compleity and Algorithms Sept 8, 2015 Lecture 4: LMN Learning (Part 2) Instructor: Russell Impagliazzo Scribe: Preetum Nakkiran 1 Overview Continuing from last lecture, we will
More informationLecture 5: February 21, 2017
COMS 6998: Advanced Complexity Spring 2017 Lecture 5: February 21, 2017 Lecturer: Rocco Servedio Scribe: Diana Yin 1 Introduction 1.1 Last Time 1. Established the exact correspondence between communication
More informationOn the correlation of parity and small-depth circuits
Electronic Colloquium on Computational Complexity, Report No. 137 (2012) On the correlation of parity and small-depth circuits Johan Håstad KTH - Royal Institute of Technology November 1, 2012 Abstract
More informationLecture 10: Learning DNF, AC 0, Juntas. 1 Learning DNF in Almost Polynomial Time
Analysis of Boolean Functions (CMU 8-859S, Spring 2007) Lecture 0: Learning DNF, AC 0, Juntas Feb 5, 2007 Lecturer: Ryan O Donnell Scribe: Elaine Shi Learning DNF in Almost Polynomial Time From previous
More informationGuest Lecture: Average Case depth hierarchy for,, -Boolean circuits. Boaz Barak
Guest Lecture: Average Case depth hierarchy for,, -Boolean circuits. Boaz Barak Thanks to the authors of RST for answering my questions by email and in particular to Li-Yang Tan for sharing with me his
More informationTesting by Implicit Learning
Testing by Implicit Learning Rocco Servedio Columbia University ITCS Property Testing Workshop Beijing January 2010 What this talk is about 1. Testing by Implicit Learning: method for testing classes of
More informationcomplexity distributions
The of complexity distributions Emanuele Viola Northeastern University March 2012 Local functions (a.k.a. Junta, NC 0 ) f : {0,1}n {0,1} d-local : output depends on d input bits Input x d f Fact: Parity(x)
More informationLecture 5: February 16, 2012
COMS 6253: Advanced Computational Learning Theory Lecturer: Rocco Servedio Lecture 5: February 16, 2012 Spring 2012 Scribe: Igor Carboni Oliveira 1 Last time and today Previously: Finished first unit on
More information1 AC 0 and Håstad Switching Lemma
princeton university cos 522: computational complexity Lecture 19: Circuit complexity Lecturer: Sanjeev Arora Scribe:Tony Wirth Complexity theory s Waterloo As we saw in an earlier lecture, if PH Σ P 2
More informationLearning and Fourier Analysis
Learning and Fourier Analysis Grigory Yaroslavtsev http://grigory.us Slides at http://grigory.us/cis625/lecture2.pdf CIS 625: Computational Learning Theory Fourier Analysis and Learning Powerful tool for
More informationSubspace-Invariant AC 0 Formulas
Subspace-Invariant AC 0 Formulas Benjamin Rossman University of Toronto June 12, 2018 Abstract We consider the action of a linear subspace U of {0, 1} n on the set of AC 0 formulas with inputs labeled
More informationLecture 3: AC 0, the switching lemma
Lecture 3: AC 0, the switching lemma Topics in Complexity Theory and Pseudorandomness (Spring 2013) Rutgers University Swastik Kopparty Scribes: Meng Li, Abdul Basit 1 Pseudorandom sets We start by proving
More information3 Finish learning monotone Boolean functions
COMS 6998-3: Sub-Linear Algorithms in Learning and Testing Lecturer: Rocco Servedio Lecture 5: 02/19/2014 Spring 2014 Scribes: Dimitris Paidarakis 1 Last time Finished KM algorithm; Applications of KM
More informationCS Communication Complexity: Applications and New Directions
CS 2429 - Communication Complexity: Applications and New Directions Lecturer: Toniann Pitassi 1 Introduction In this course we will define the basic two-party model of communication, as introduced in the
More information6.842 Randomness and Computation April 2, Lecture 14
6.84 Randomness and Computation April, 0 Lecture 4 Lecturer: Ronitt Rubinfeld Scribe: Aaron Sidford Review In the last class we saw an algorithm to learn a function where very little of the Fourier coeffecient
More informationLearning Sums of Independent Integer Random Variables
Learning Sums of Independent Integer Random Variables Cos8s Daskalakis (MIT) Ilias Diakonikolas (Edinburgh) Ryan O Donnell (CMU) Rocco Servedio (Columbia) Li- Yang Tan (Columbia) FOCS 2013, Berkeley CA
More informationTwo Sides of the Coin Problem
Electronic Colloquium on Computational Complexity, Report No. 23 (2014) Two Sides of the Coin Problem Gil Cohen Anat Ganor Ran Raz February 19, 2014 Abstract In the coin problem, one is given n independent
More informationNotes for Lecture 25
U.C. Berkeley CS278: Computational Complexity Handout N25 ofessor Luca Trevisan 12/1/2004 Notes for Lecture 25 Circuit Lower Bounds for Parity Using Polynomials In this lecture we prove a lower bound on
More informationSome Depth Two (and Three) Threshold Circuit Lower Bounds. Ryan Williams Stanford Joint work with Daniel Kane (UCSD)
Some Depth Two (and Three) Threshold Circuit Lower Bounds Ryan Williams Stanford Joint wor with Daniel Kane (UCSD) Introduction Def. f n : 0,1 n 0,1 is a linear threshold function (LTF) if there are w
More informationLinear sketching for Functions over Boolean Hypercube
Linear sketching for Functions over Boolean Hypercube Grigory Yaroslavtsev (Indiana University, Bloomington) http://grigory.us with Sampath Kannan (U. Pennsylvania), Elchanan Mossel (MIT) and Swagato Sanyal
More informationCertifying polynomials for AC 0 [ ] circuits, with applications
Certifying polynomials for AC 0 [ ] circuits, with applications Swastik Kopparty Srikanth Srinivasan Abstract In this paper, we introduce and develop the method of certifying polynomials for proving AC
More informationCS Introduction to Complexity Theory. Lecture #11: Dec 8th, 2015
CS 2401 - Introduction to Complexity Theory Lecture #11: Dec 8th, 2015 Lecturer: Toniann Pitassi Scribe Notes by: Xu Zhao 1 Communication Complexity Applications Communication Complexity (CC) has many
More informationLearning and Fourier Analysis
Learning and Fourier Analysis Grigory Yaroslavtsev http://grigory.us CIS 625: Computational Learning Theory Fourier Analysis and Learning Powerful tool for PAC-style learning under uniform distribution
More informationComplexity Theory of Polynomial-Time Problems
Complexity Theory of Polynomial-Time Problems Lecture 3: The polynomial method Part I: Orthogonal Vectors Sebastian Krinninger Organization of lecture No lecture on 26.05. (State holiday) 2 nd exercise
More informationTesting Monotone High-Dimensional Distributions
Testing Monotone High-Dimensional Distributions Ronitt Rubinfeld Computer Science & Artificial Intelligence Lab. MIT Cambridge, MA 02139 ronitt@theory.lcs.mit.edu Rocco A. Servedio Department of Computer
More informationRandomized Complexity Classes; RP
Randomized Complexity Classes; RP Let N be a polynomial-time precise NTM that runs in time p(n) and has 2 nondeterministic choices at each step. N is a polynomial Monte Carlo Turing machine for a language
More informationTight bounds on The Fourier Spectrum of AC 0
Electronic Colloquium on Computational Complexity, Report No. 174 (2014) Tight bounds on The Fourier Spectrum of AC 0 Avishay Tal December 14, 2014 Abstract We show that AC 0 circuits of depth d and size
More information5.1 Learning using Polynomial Threshold Functions
CS 395T Computational Learning Theory Lecture 5: September 17, 2007 Lecturer: Adam Klivans Scribe: Aparajit Raghavan 5.1 Learning using Polynomial Threshold Functions 5.1.1 Recap Definition 1 A function
More informationarxiv: v1 [cs.cc] 13 Feb 2017
Separation of AC 0 [ ] Formulas and Circuits arxiv:1702.03625v1 [cs.cc] 13 Feb 2017 Benjamin Rossman Depts. of Math and Computer Science University of Toronto February 14, 2017 Abstract Srikanth Srinivasan
More informationLecture 29: Computational Learning Theory
CS 710: Complexity Theory 5/4/2010 Lecture 29: Computational Learning Theory Instructor: Dieter van Melkebeek Scribe: Dmitri Svetlov and Jake Rosin Today we will provide a brief introduction to computational
More informationCS 395T Computational Learning Theory. Scribe: Mike Halcrow. x 4. x 2. x 6
CS 395T Computational Learning Theory Lecture 3: September 0, 2007 Lecturer: Adam Klivans Scribe: Mike Halcrow 3. Decision List Recap In the last class, we determined that, when learning a t-decision list,
More informationCommunication is bounded by root of rank
Electronic Colloquium on Computational Complexity, Report No. 84 (2013) Communication is bounded by root of rank Shachar Lovett June 7, 2013 Abstract We prove that any total boolean function of rank r
More informationNotes for Lecture 11
Stanford University CS254: Computational Complexity Notes 11 Luca Trevisan 2/11/2014 Notes for Lecture 11 Circuit Lower Bounds for Parity Using Polynomials In this lecture we prove a lower bound on the
More information1 Randomized reduction: a first start
6.S078 Fine-Grained Algorithms and Complexity MIT Lectures 3 and 4 February 14 and 21, 2018 Today: In the previous two lectures, we mentioned ETH and SETH, about the time complexity of SAT. Today we want
More informationMeta-Algorithms vs. Circuit Lower Bounds Valentine Kabanets
Meta-Algorithms vs. Circuit Lower Bounds Valentine Kabanets Tokyo Institute of Technology & Simon Fraser University Understanding Efficiency Better understanding of Efficient Computation Good Algorithms
More informationAn Improved Deterministic #SAT Algorithm for Small De Morgan Formulas
Noname manuscript No. (will be inserted by the editor) An Improved Deterministic #SAT Algorithm for Small De Morgan Formulas Ruiwen Chen Valentine Kabanets Nitin Saurabh Received: date / Accepted: date
More informationc 2004 Society for Industrial and Applied Mathematics
SIAM J. COMPUT. Vol. 34, No. 2, pp. 261 276 c 2004 Society for Industrial and Applied Mathematics BOUNDED-DEPTH FREGE LOWER BOUNDS FOR WEAKER PIGEONHOLE PRINCIPLES JOSHUA BURESH-OPPENHEIM, PAUL BEAME,
More informationOn converting CNF to DNF
Electronic Colloquium on Computational Complexity, Report No. 17 (2003) On converting CNF to DNF Peter Bro Miltersen Jaikumar Radhakrishnan Ingo Wegener March 27, 2003 Abstract The best-known representations
More informationMulti-Linear Formulas for Permanent and Determinant are of Super-Polynomial Size
Multi-Linear Formulas for Permanent and Determinant are of Super-Polynomial Size Ran Raz Weizmann Institute ranraz@wisdom.weizmann.ac.il Abstract An arithmetic formula is multi-linear if the polynomial
More informationComplete problems for classes in PH, The Polynomial-Time Hierarchy (PH) oracle is like a subroutine, or function in
Oracle Turing Machines Nondeterministic OTM defined in the same way (transition relation, rather than function) oracle is like a subroutine, or function in your favorite PL but each call counts as single
More informationThe Analysis of Partially Symmetric Functions
The Analysis of Partially Symmetric Functions Eric Blais Based on joint work with Amit Weinstein Yuichi Yoshida Classes of simple functions Classes of simple functions Constant Classes of simple functions
More informationCS Foundations of Communication Complexity
CS 49 - Foundations of Communication Complexity Lecturer: Toniann Pitassi 1 The Discrepancy Method Cont d In the previous lecture we ve outlined the discrepancy method, which is a method for getting lower
More informationLecture 3: Nondeterminism, NP, and NP-completeness
CSE 531: Computational Complexity I Winter 2016 Lecture 3: Nondeterminism, NP, and NP-completeness January 13, 2016 Lecturer: Paul Beame Scribe: Paul Beame 1 Nondeterminism and NP Recall the definition
More informationStructure of protocols for XOR functions
Electronic Colloquium on Computational Complexity, Report No. 44 (016) Structure of protocols for XOR functions Kaave Hosseini Computer Science and Engineering University of California, San Diego skhossei@ucsd.edu
More informationLecture Notes 3 Convergence (Chapter 5)
Lecture Notes 3 Convergence (Chapter 5) 1 Convergence of Random Variables Let X 1, X 2,... be a sequence of random variables and let X be another random variable. Let F n denote the cdf of X n and let
More informationCMPUT 675: Approximation Algorithms Fall 2014
CMPUT 675: Approximation Algorithms Fall 204 Lecture 25 (Nov 3 & 5): Group Steiner Tree Lecturer: Zachary Friggstad Scribe: Zachary Friggstad 25. Group Steiner Tree In this problem, we are given a graph
More informationUpper Bounds on Fourier Entropy
Upper Bounds on Fourier Entropy Sourav Chakraborty 1, Raghav Kulkarni 2, Satyanarayana V. Lokam 3, and Nitin Saurabh 4 1 Chennai Mathematical Institute, Chennai, India sourav@cmi.ac.in 2 Centre for Quantum
More informationThe Strength of Multilinear Proofs
The Strength of Multilinear Proofs Ran Raz Iddo Tzameret December 19, 2006 Abstract We introduce an algebraic proof system that manipulates multilinear arithmetic formulas. We show that this proof system
More informationRegular Resolution Lower Bounds for the Weak Pigeonhole Principle
Regular Resolution Lower Bounds for the Weak Pigeonhole Principle Toniann Pitassi Department of Computer Science University of Toronto toni@cs.toronto.edu Ran Raz Department of Computer Science Weizmann
More informationImproved Direct Product Theorems for Randomized Query Complexity
Improved Direct Product Theorems for Randomized Query Complexity Andrew Drucker Nov. 16, 2010 Andrew Drucker, Improved Direct Product Theorems for Randomized Query Complexity 1/28 Big picture Usually,
More informationJunta Approximations for Submodular, XOS and Self-Bounding Functions
Junta Approximations for Submodular, XOS and Self-Bounding Functions Vitaly Feldman Jan Vondrák IBM Almaden Research Center Simons Institute, Berkeley, October 2013 Feldman-Vondrák Approximations by Juntas
More informationCS 151 Complexity Theory Spring Solution Set 5
CS 151 Complexity Theory Spring 2017 Solution Set 5 Posted: May 17 Chris Umans 1. We are given a Boolean circuit C on n variables x 1, x 2,..., x n with m, and gates. Our 3-CNF formula will have m auxiliary
More informationAverage-Case Lower Bounds and Satisfiability Algorithms for Small Threshold Circuits
Average-Case Lower Bounds and Satisfiability Algorithms for Small Threshold Circuits Ruiwen Chen 1, Rahul Santhanam 2, and Srikanth Srinivasan 3 1 University of Oxford, Oxford, UK ruiwen.chen@cs.ox.ac.uk
More informationEfficient Approximation for Restricted Biclique Cover Problems
algorithms Article Efficient Approximation for Restricted Biclique Cover Problems Alessandro Epasto 1, *, and Eli Upfal 2 ID 1 Google Research, New York, NY 10011, USA 2 Department of Computer Science,
More informationGraphical Models. Lecture 10: Variable Elimina:on, con:nued. Andrew McCallum
Graphical Models Lecture 10: Variable Elimina:on, con:nued Andrew McCallum mccallum@cs.umass.edu Thanks to Noah Smith and Carlos Guestrin for some slide materials. 1 Last Time Probabilis:c inference is
More informationEvaluation of DNF Formulas
Evaluation of DNF Formulas Sarah R. Allen Carnegie Mellon University Computer Science Department Pittsburgh, PA, USA Lisa Hellerstein and Devorah Kletenik Polytechnic Institute of NYU Dept. of Computer
More informationAlternating nonzero automata
Alternating nonzero automata Application to the satisfiability of CTL [,, P >0, P =1 ] Hugo Gimbert, joint work with Paulin Fournier LaBRI, Université de Bordeaux ANR Stoch-MC 06/07/2017 Control and verification
More informationASwitchingLemmaPrimer
ASwitchingLemmaPrimer Paul Beame University of Toronto 10 King s College Road Toronto, Ontario Canada M5S 1A4 beame@cs.washington.edu April 94 Abstract We present simplified proofs for a variety of switching
More informationCSE 21 Math for Algorithms and Systems Analysis. Lecture 10 Condi<onal Probability
CSE 21 Math for Algorithms and Systems Analysis Lecture 10 Condi
More informationIST 4 Information and Logic
IST 4 Information and Logic mon tue wed thr fri sun T = today 3 M oh x= hw#x out 0 oh M 7 oh oh 2 M2 oh oh x= hw#x due 24 oh oh 2 oh = office hours oh oh M2 8 3 oh midterms oh oh Mx= MQx out 5 oh 3 4 oh
More informationExpectation. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda
Expectation DS GA 1002 Statistical and Mathematical Models http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall16 Carlos Fernandez-Granda Aim Describe random variables with a few numbers: mean, variance,
More informationOn the Fourier spectrum of symmetric Boolean functions
On the Fourier spectrum of symmetric Boolean functions Amir Shpilka Technion and MSR NE Based on joint work with Avishay Tal 1 Theme: Analysis of Boolean functions Pick favorite representation: Fourier
More informationPoly-logarithmic independence fools AC 0 circuits
Poly-logarithmic independence fools AC 0 circuits Mark Braverman Microsoft Research New England January 30, 2009 Abstract We prove that poly-sized AC 0 circuits cannot distinguish a poly-logarithmically
More informationThe Monte Carlo Method
The Monte Carlo Method Example: estimate the value of π. Choose X and Y independently and uniformly at random in [0, 1]. Let Pr(Z = 1) = π 4. 4E[Z] = π. { 1 if X Z = 2 + Y 2 1, 0 otherwise, Let Z 1,...,
More informationLecture 24: Randomized Complexity, Course Summary
6.045 Lecture 24: Randomized Complexity, Course Summary 1 1/4 1/16 1/4 1/4 1/32 1/16 1/32 Probabilistic TMs 1/16 A probabilistic TM M is a nondeterministic TM where: Each nondeterministic step is called
More informationEvolutionary Dynamics on Graphs
Evolutionary Dynamics on Graphs Leslie Ann Goldberg, University of Oxford Absorption Time of the Moran Process (2014) with Josep Díaz, David Richerby and Maria Serna Approximating Fixation Probabilities
More informationLow degree almost Boolean functions are sparse juntas
Low degree almost Boolean functions are sparse juntas Irit Dinur Yuval Filmus Prahladh Harsha November 19, 2017 Abstract Nisan and Szegedy showed that low degree Boolean functions are juntas. Kindler and
More informationLower Bounds for Cutting Planes Proofs. with Small Coecients. Abstract. We consider small-weight Cutting Planes (CP ) proofs; that is,
Lower Bounds for Cutting Planes Proofs with Small Coecients Maria Bonet y Toniann Pitassi z Ran Raz x Abstract We consider small-weight Cutting Planes (CP ) proofs; that is, Cutting Planes (CP ) proofs
More informationLecture 16: Communication Complexity
CSE 531: Computational Complexity I Winter 2016 Lecture 16: Communication Complexity Mar 2, 2016 Lecturer: Paul Beame Scribe: Paul Beame 1 Communication Complexity In (two-party) communication complexity
More informationQuantum algorithms based on span programs
Quantum algorithms based on span programs Ben Reichardt IQC, U Waterloo [arxiv:0904.2759] Model: Complexity measure: Quantum algorithms Black-box query complexity Span programs Witness size [KW 93] [RŠ
More informationResolution Lower Bounds for the Weak Pigeonhole Principle
Electronic Colloquium on Computational Complexity, Report No. 21 (2001) Resolution Lower Bounds for the Weak Pigeonhole Principle Ran Raz Weizmann Institute, and The Institute for Advanced Study ranraz@wisdom.weizmann.ac.il
More informationClique is hard on average for regular resolution
Clique is hard on average for regular resolution Ilario Bonacina, UPC Barcelona Tech July 20, 2018 RaTLoCC, Bertinoro How hard is to certify that a graph is Ramsey? Ilario Bonacina, UPC Barcelona Tech
More informationFPT hardness for Clique and Set Cover with super exponential time in k
FPT hardness for Clique and Set Cover with super exponential time in k Mohammad T. Hajiaghayi Rohit Khandekar Guy Kortsarz Abstract We give FPT-hardness for setcover and clique with super exponential time
More informationMajority is incompressible by AC 0 [p] circuits
Majority is incompressible by AC 0 [p] circuits Igor Carboni Oliveira Columbia University Joint work with Rahul Santhanam (Univ. Edinburgh) 1 Part 1 Background, Examples, and Motivation 2 Basic Definitions
More informationActive Testing! Liu Yang! Joint work with Nina Balcan,! Eric Blais and Avrim Blum! Liu Yang 2011
! Active Testing! Liu Yang! Joint work with Nina Balcan,! Eric Blais and Avrim Blum! 1 Property Testing Instance space X = R n (Distri D over X)! Tested function f : X->{0,1}! A property P of Boolean fn
More informationLecture 4. 1 Circuit Complexity. Notes on Complexity Theory: Fall 2005 Last updated: September, Jonathan Katz
Notes on Complexity Theory: Fall 2005 Last updated: September, 2005 Jonathan Katz Lecture 4 1 Circuit Complexity Circuits are directed, acyclic graphs where nodes are called gates and edges are called
More informationA Noisy-Influence Regularity Lemma for Boolean Functions Chris Jones
A Noisy-Influence Regularity Lemma for Boolean Functions Chris Jones Abstract We present a regularity lemma for Boolean functions f : {, } n {, } based on noisy influence, a measure of how locally correlated
More informationCOMP 562: Introduction to Machine Learning
COMP 562: Introduction to Machine Learning Lecture 20 : Support Vector Machines, Kernels Mahmoud Mostapha 1 Department of Computer Science University of North Carolina at Chapel Hill mahmoudm@cs.unc.edu
More informationDistributed Systems Gossip Algorithms
Distributed Systems Gossip Algorithms He Sun School of Informatics University of Edinburgh What is Gossip? Gossip algorithms In a gossip algorithm, each node in the network periodically exchanges information
More informationApproximation Algorithms and Hardness of Approximation. IPM, Jan Mohammad R. Salavatipour Department of Computing Science University of Alberta
Approximation Algorithms and Hardness of Approximation IPM, Jan 2006 Mohammad R. Salavatipour Department of Computing Science University of Alberta 1 Introduction For NP-hard optimization problems, we
More information6.895 Randomness and Computation March 19, Lecture Last Lecture: Boosting Weak Learners Into Strong Learners
6.895 Randomness and Computation March 9, 2008 Lecture 3 Lecturer: Ronitt Rubinfeld Scribe: Edwin Chen Overview. Last Lecture: Boosting Weak Learners Into Strong Learners In the last two lectures, we showed
More informationCSC 5170: Theory of Computational Complexity Lecture 9 The Chinese University of Hong Kong 15 March 2010
CSC 5170: Theory of Computational Complexity Lecture 9 The Chinese University of Hong Kong 15 March 2010 We now embark on a study of computational classes that are more general than NP. As these classes
More informationCircuits. Lecture 11 Uniform Circuit Complexity
Circuits Lecture 11 Uniform Circuit Complexity 1 Recall 2 Recall Non-uniform complexity 2 Recall Non-uniform complexity P/1 Decidable 2 Recall Non-uniform complexity P/1 Decidable NP P/log NP = P 2 Recall
More informationACO Comprehensive Exam March 20 and 21, Computability, Complexity and Algorithms
1. Computability, Complexity and Algorithms Part a: You are given a graph G = (V,E) with edge weights w(e) > 0 for e E. You are also given a minimum cost spanning tree (MST) T. For one particular edge
More informationMachine Learning
Machine Learning 10-601 Tom M. Mitchell Machine Learning Department Carnegie Mellon University August 30, 2017 Today: Decision trees Overfitting The Big Picture Coming soon Probabilistic learning MLE,
More informationBinary Decision Diagrams Boolean Functions
Binary Decision Diagrams Representation of Boolean Functions BDDs, OBDDs, ROBDDs Operations Model-Checking over BDDs 72 Boolean functions:b = {0,1}, f :B B B Boolean Functions Boolean expressions: t ::=
More informationPolynomial time Prediction Strategy with almost Optimal Mistake Probability
Polynomial time Prediction Strategy with almost Optimal Mistake Probability Nader H. Bshouty Department of Computer Science Technion, 32000 Haifa, Israel bshouty@cs.technion.ac.il Abstract We give the
More informationClique is hard on average for regular resolution
Clique is hard on average for regular resolution Ilario Bonacina, UPC Barcelona Tech July 27, 2018 Oxford Complexity Day Talk based on a joint work with: A. Atserias S. de Rezende J. Nordstro m M. Lauria
More informationLower Bounds for Bounded Depth Frege Proofs via Buss-Pudlák Games
Lower Bounds for Bounded Depth Frege Proofs via Buss-Pudlák Games Eli Ben-Sasson Prahladh Harsha Abstract We present a simple proof of the bounded-depth Frege lower bounds of Pitassi et. al. and Krajíček
More informationUpper and Lower Bounds for Tree-like. Cutting Planes Proofs. Abstract. In this paper we study the complexity of Cutting Planes (CP) refutations, and
Upper and Lower Bounds for Tree-like Cutting Planes Proofs Russell Impagliazzo Toniann Pitassi y Alasdair Urquhart z UCSD UCSD and U. of Pittsburgh University of Toronto Abstract In this paper we study
More informationDegree and Sensitivity: tails of two distributions *
Electronic Colloquium on Computational Complexity, Report No. 69 (2016) Degree and Sensitivity: tails of two distributions * Parikshit Gopalan Microsoft Research parik@microsoft.com Rocco A. Servedio Columbia
More informationBias/variance tradeoff, Model assessment and selec+on
Applied induc+ve learning Bias/variance tradeoff, Model assessment and selec+on Pierre Geurts Department of Electrical Engineering and Computer Science University of Liège October 29, 2012 1 Supervised
More information25 Minimum bandwidth: Approximation via volume respecting embeddings
25 Minimum bandwidth: Approximation via volume respecting embeddings We continue the study of Volume respecting embeddings. In the last lecture, we motivated the use of volume respecting embeddings by
More informationAn Introduction to Randomized algorithms
An Introduction to Randomized algorithms C.R. Subramanian The Institute of Mathematical Sciences, Chennai. Expository talk presented at the Research Promotion Workshop on Introduction to Geometric and
More informationQuantum boolean functions
Quantum boolean functions Ashley Montanaro 1 and Tobias Osborne 2 1 Department of Computer Science 2 Department of Mathematics University of Bristol Royal Holloway, University of London Bristol, UK London,
More information