Communication Complexity 16:198: February Lecture 4. x ij y ij

Similar documents
princeton univ. F 17 cos 521: Advanced Algorithm Design Lecture 7: LP Duality Lecturer: Matt Weinberg

Assortment Optimization under MNL

Matrix Approximation via Sampling, Subspace Embedding. 1 Solving Linear Systems Using SVD

Lectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix

COS 521: Advanced Algorithms Game Theory and Linear Programming

Stanford University CS359G: Graph Partitioning and Expanders Handout 4 Luca Trevisan January 13, 2011

Errors for Linear Systems

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems

Affine transformations and convexity

U.C. Berkeley CS294: Beyond Worst-Case Analysis Luca Trevisan September 5, 2017

6.842 Randomness and Computation February 18, Lecture 4

Lecture 3. Ax x i a i. i i

Lecture 20: Lift and Project, SDP Duality. Today we will study the Lift and Project method. Then we will prove the SDP duality theorem.

Lecture 4: Constant Time SVD Approximation

Yong Joon Ryang. 1. Introduction Consider the multicommodity transportation problem with convex quadratic cost function. 1 2 (x x0 ) T Q(x x 0 )

Lecture 10: May 6, 2013

Solutions to exam in SF1811 Optimization, Jan 14, 2015

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results.

Math 217 Fall 2013 Homework 2 Solutions

Maximizing the number of nonnegative subsets

PHYS 705: Classical Mechanics. Calculus of Variations II

Lecture 10 Support Vector Machines II

College of Computer & Information Science Fall 2009 Northeastern University 20 October 2009

Lecture 17: Lee-Sidford Barrier

APPENDIX A Some Linear Algebra

Min Cut, Fast Cut, Polynomial Identities

CS : Algorithms and Uncertainty Lecture 17 Date: October 26, 2016

Some basic inequalities. Definition. Let V be a vector space over the complex numbers. An inner product is given by a function, V V C

Module 9. Lecture 6. Duality in Assignment Problems

Singular Value Decomposition: Theory and Applications

Linear, affine, and convex sets and hulls In the sequel, unless otherwise specified, X will denote a real vector space.

BOUNDEDNESS OF THE RIESZ TRANSFORM WITH MATRIX A 2 WEIGHTS

Lecture 3: Dual problems and Kernels

U.C. Berkeley CS294: Spectral Methods and Expanders Handout 8 Luca Trevisan February 17, 2016

8.4 COMPLEX VECTOR SPACES AND INNER PRODUCTS

10-701/ Machine Learning, Fall 2005 Homework 3

The Second Anti-Mathima on Game Theory

Lecture Notes on Linear Regression

Grover s Algorithm + Quantum Zeno Effect + Vaidman

The Order Relation and Trace Inequalities for. Hermitian Operators

Lecture 12: Discrete Laplacian

10-801: Advanced Optimization and Randomized Methods Lecture 2: Convex functions (Jan 15, 2014)

Difference Equations

Finding Dense Subgraphs in G(n, 1/2)

Homework Notes Week 7

Dynamic Programming. Preview. Dynamic Programming. Dynamic Programming. Dynamic Programming (Example: Fibonacci Sequence)

Integrals and Invariants of Euler-Lagrange Equations

n α j x j = 0 j=1 has a nontrivial solution. Here A is the n k matrix whose jth column is the vector for all t j=0

SELECTED SOLUTIONS, SECTION (Weak duality) Prove that the primal and dual values p and d defined by equations (4.3.2) and (4.3.3) satisfy p d.

Math 426: Probability MWF 1pm, Gasson 310 Homework 4 Selected Solutions

DECOUPLING THEORY HW2

σ τ τ τ σ τ τ τ σ Review Chapter Four States of Stress Part Three Review Review

Lecture 2: Gram-Schmidt Vectors and the LLL Algorithm

Edge Isoperimetric Inequalities

Calculation of time complexity (3%)

Estimation: Part 2. Chapter GREG estimation

Some modelling aspects for the Matlab implementation of MMA

The lower and upper bounds on Perron root of nonnegative irreducible matrices

THE ARIMOTO-BLAHUT ALGORITHM FOR COMPUTATION OF CHANNEL CAPACITY. William A. Pearlman. References: S. Arimoto - IEEE Trans. Inform. Thy., Jan.

C/CS/Phy191 Problem Set 3 Solutions Out: Oct 1, 2008., where ( 00. ), so the overall state of the system is ) ( ( ( ( 00 ± 11 ), Φ ± = 1

( ) 2 ( ) ( ) Problem Set 4 Suggested Solutions. Problem 1

Exercise Solutions to Real Analysis

Generalized Linear Methods

Notes on Frequency Estimation in Data Streams

Expected Value and Variance

Lecture 21: Numerical methods for pricing American type derivatives

THE CHINESE REMAINDER THEOREM. We should thank the Chinese for their wonderful remainder theorem. Glenn Stevens

REAL ANALYSIS I HOMEWORK 1

Feature Selection: Part 1

MMA and GCMMA two methods for nonlinear optimization

1 Convex Optimization

Inner Product. Euclidean Space. Orthonormal Basis. Orthogonal

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)

Maximal Margin Classifier

Learning Theory: Lecture Notes

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 12 10/21/2013. Martingale Concentration Inequalities and Applications

The Minimum Universal Cost Flow in an Infeasible Flow Network

18.1 Introduction and Recap

P exp(tx) = 1 + t 2k M 2k. k N

Problem Do any of the following determine homomorphisms from GL n (C) to GL n (C)?

Foundations of Arithmetic

MEM 255 Introduction to Control Systems Review: Basics of Linear Algebra

MAT 578 Functional Analysis

763622S ADVANCED QUANTUM MECHANICS Solution Set 1 Spring c n a n. c n 2 = 1.

p 1 c 2 + p 2 c 2 + p 3 c p m c 2

Stanford University CS254: Computational Complexity Notes 7 Luca Trevisan January 29, Notes for Lecture 7

LOW BIAS INTEGRATED PATH ESTIMATORS. James M. Calvin

Société de Calcul Mathématique SA

Lecture 5 September 17, 2015

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U)

A Local Variational Problem of Second Order for a Class of Optimal Control Problems with Nonsmooth Objective Function

Canonical transformations

6) Derivatives, gradients and Hessian matrices

NP-Completeness : Proofs

CS286r Assign One. Answer Key

The Geometry of Logit and Probit

LECTURE 9 CANONICAL CORRELATION ANALYSIS

9 Characteristic classes

Lecture 10 Support Vector Machines. Oct

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Transcription:

Communcaton Complexty 16:198:671 09 February 2010 Lecture 4 Lecturer: Troy Lee Scrbe: Rajat Mttal 1 Homework problem : Trbes We wll solve the thrd queston n the homework. The goal s to show that the nondetermnstc complexty (N 1 (f), N 0 (f)) of TRIES n functon s O( n) and determnstc complexty s Ω(n). Hence the gap between N 1 (f), N 0 (f) and D(f) s essentally tght. The trbes functon s gven by TRIES n (x, y) = n n =1 j=1 x j y j Let s calculate the nondetermnstc complexty N 1 (f) frst. It s enough to specfy postons n all the n clauses where both x j, y j take the value one. It takes (log n) bts to specfy a poston n n long clause. Snce there are n such clauses N 1 (f) n log n = O( n log n). In the artcle [1] an upper bound of O( n) s clamed. The log factor, however, s actually requred. One can create a foolng set F = {(x, x) : x has exactly one one n each block}. There are n n many such nputs and they all satsfy TRIES n (x, x) = 1. For (x, x), (x, x ) F where x x, however, we have TRIES n (x, x ) = 0. One can show that N 0 (f) = O( n). In case the functon evaluates to zero the prover can gve the ndex of an nput to of the AND gate whch s not satsfed wth log( n) bts, and then the players can verfy wth O( n) bts of communcaton that ther nputs do not ntersect on ths block. For the determnstc complexty lower bound, we wll use tensor product trck. Frst notce that every clause s solvng a set ntersecton problem..e., gven two strngs, fnd f ther s a common ndex (poston) where both are one. Set ntersecton problem s the negaton of the dsjontness problem. The dsjontness problem s DISJ n = n =1 (x y ) The matrx for DISJ 1, the dsjontness functon on one bt, s gven by 1 1 1 0 We clam that DISJ n = n =1DISJ 1. Gven two matrces A,, ther tensor product s (A )[(x 1, x 2 ), (y 1, y 2 )] = A(x 1, y 1 ) (x 2, y 2 ) 1

So when the matrces have entres from {0, 1}, the tensor product A expresses the AND of all pars from A and. Two n-bt strngs x, y are dsjont f and only f x 1,..., x n/2, y 1,..., y n/2 are dsjont AND x n/2 +1,..., x n, y n/2 +1,..., y n are dsjont, we see that by expandng ths expresson recursvely that DISJ) n = n =1DISJ 1. Also for the tensor product we know rk(a ) = rk(a) rk(). Notce that the matrx for DISJ 1 s full rank, so the rank of DISJ n s 2 n. Snce set ntersecton s the negaton of dsjontness problem, DISJ n = J SI n, where J s the all 1 s matrx. y the subaddtvty of rank rk(disj n ) rk(j) + rk(si n ) rk(si n ) 2 n 1 The TRIES n functon s the AND of n many set ntersecton problems. Thus the communcaton matrx for TRIES n s the tensor product of n copes of the communcaton matrx for SI n. Agan by multplcatvty of rank under tensor product we have. rk(tries n (x, y)) (2 n 1) n, and so D(TRIES n ) log(rk(tries n )) = Ω(n). 2 Randomzed Communcaton Complexty In the last lecture, we defned randomzed communcaton complexty. There are two versons, publc con and prvate con. We wll focus on publc con randomzed complexty (recall that by Newman s theorem the two models are equvalent, up to an addtve logarthmc factor). A publc con protocol s a probablty dstrbuton over determnstc protocols {P r }. A protocol has error at most ɛ f Pr r [P r (x, y) = f(x, y)] 1 ɛ for all (x, y). The cost of a randomzed protocol s max r D(P r ). As usual the randomzed complexty R ɛ (f) s the mnmum cost of a protocol whch computes f wth error at most ɛ. 2.1 Lower bounds on randomzed communcaton complexty Consder a randomzed protocol P whch wth probablty p r runs the determnstc protocol P r. We dentfy P r wth the oolean matrx whose (x, y) entry s the output of P r on (x, y). Let M P be the matrx whose (x, y) entry represents the probablty that the protocol outputs 1. In other words, M P = r p rp r. Consder a functon wth communcaton matrx F. If the protocol P computes F wth error at most ɛ, then we have that F M P ɛ. The fact that a randomzed protocol s a convex combnaton of determnstc protocols allows us to qute generally transform a lower bound technque for determnstc complexty wth nce propertes nto a lower bound for randomzed complexty. Theorem 1. Suppose there exst a functon Φ : X Y R, whch satsfes the followng propertes 1. Φ(F ) C P (F ) 2

2. Φ(x + y) Φ(x) + Φ(y) 3. Φ(cx) = c Φ(x) Let Φ ɛ (F ) = mn G {Φ(G) : G F ɛ} (remember ths notaton, we wll use t for dfferent functons also). Then R ɛ (F ) log Φ ɛ (F ) Proof : The matrx for protocol P s ɛ away from the matrx for F. So we get Φ ɛ (F ) Φ(M P ). Now we compute Φ(M P ). Φ(M P ) = Φ( r p r P r ) p r Φ(P r ) p r C P (P r ) p r 2 D(Pr) p r 2 D(Pr ) (where D(P r ) s maxmum for r*) 2 D(Pr ) 2 Rɛ(F ) So we get Φ ɛ (F ) 2 Rɛ(F ). In other words R ɛ (F ) log Φ ɛ (F ). 2.2 Examples for Theorem 1 Lets take example of rk(f ) and see f t fts n the framework of Theorem 1. We know rk(f ) C P (F ) and also that rk(x+y) rk(x)+rk(y). ut the thrd property doesn t work because rk(cx) = rk(x). Lets look at rank as an optmzaton program and try to relax t to some other combnatoral quantty whch wll ft nto theorem s framework. Notce that C D (F ) = mn α such that F = α x y t where x, y {0, 1} 2n α {0, 1} 3

Rank can be wrtten as a relaxaton of ths program rk(f ) = mn α such that F = α x y t where x, y R 2n α {0, 1} We see that rank relaxes the constrant that the x, y are oolean vectors and allows them to be real vectors, but keeps the constrant that α {0, 1}, a quadratc constrant. Let us relax C D (F ) n a dfferent way. µ(f ) = mn α such that F = α x y t where x, y {0, 1} 2n α R Here we keep x, y to be oolean vectors, but relax the constrant on α. Ths quantty can now be wrtten as a lnear program, albet a huge one as there are 2 2n+1 many varables, one for each rectangle. It s easy to verfy that µ satsfes the propertes of theorem 1 and hence R ɛ (F ) log µ ɛ (F ). Whle not obvous from the defnton, t actually turns out µ(f ) 2 = O(rk(F )) for a oolean matrx F. We can even consder a tghter relaxaton of C D. Call t µ +. µ + (F ) = mn α such that F = α x y t where x, y {0, 1} 2n α 0. Notce that µ + s only defned for nonnegatve matrces (otherwse we can thnk of t as takng the value + ). ecause of ths t does not strctly satsfy condton (3) from the Theorem 1. However, t does satsfy ths for nonnegatve constants c and one can easly see that ths s all that s needed for the theorem to go through, thus we also have R ɛ (F ) log µ ɛ +(F ). 2.3 Dual of a norm Any functon Φ : X Y R whch satsfes these two condtons s called a semnorm. 1. Φ(A + ) Φ(A) + Φ() 2. Φ(cA) = c Φ(A) 4

Notce the smlarty wth condtons n Theorem 1. The dual of a semnorm Φ s defned as Φ (A) = max A, Φ() = max A,. Φ()=1 Here A, s the nner product of A, thought of as long vectors. We can use these norms to gve lower bound on randomzed communcaton complexty. Example of norms and dual norms : l 1 (v) = v. The dual norm s l 1(v) = max P u =1 v u = l (v) (the best u s the one whch puts all ts weght on the coordnate of v of largest magntude). l 2 (v) = v2 l 2(v) = l 2 (v). In general l p s dual to l q f 1 p + 1 q = 1. Now consder the dual norm of µ. Ths norm s defned as µ (A) = max A,. :µ()=1 If µ() = 1, then we can express = α R where α = 1 and each R s a rank one oolean matrx. Then we see that A, α R α A, R max A, R. Ths last value s achevable by settng α = 1 where R realzes ths maxmum. Thus we see that µ (A) = max A, R. Expressng Φ n terms of ts dual norm can be useful for showng lower bounds as t s phrased as a maxmzaton problem. Ths to lower bound Φ(A) we just have to exhbt a vector whch has non-neglgble nner product wth A and small dual norm. For ths framework we wll also need the fact that the dual of a dual norm s agan the orgnal norm. Theorem 2. Φ (v) = Φ(v) for any norm Φ. For example, ths theorem gves that µ(a) = max A, µ (). We wll use ths formulaton next to show a lower bound on the randomzed complexty of nner product. 5

3 Approxmate norms From Theorem 1 the dea of an approxmate norm naturally arose. Recall that Φ ɛ (A) = mn{φ() : A ɛ}. Note that Φ ɛ s not tself a norm n general. Let us see how we can lower bound Φ ɛ. We have Φ ɛ (A) = mn max A +, : ɛ Φ () max A, ɛl 1 () Φ () In fact t turns out that Φ ɛ s equal to ths expresson, as we wll dscuss n the next lecture. Ths means that we do not lose any power by showng a lower bound n ths fashon. Next we wll use ths expresson wth the norm µ to see a lower bound on the nner product functon. 3.1 Lower bound on Inner Product The nner product functon s defned as The matrx for IP 1 s gven by IP n (x, y) = ( 1) x,y 1 1 1-1 Ths s the famlar 2-by-2 Hadamard matrx, an orthogonal matrx (.e. has orthogonal rows) wth entres from { 1, +1}. We can agan use the tensor product to express the communcaton matrx for nner product. As we are workng over { 1, +1} the tensor product now takes the xor of the respectve entres. Ths means that IP n = IP n 1 and s also orthogonal. Now we want to lower bound µ 2ɛ (IP n ) max IP n, 2ɛl 1 (). µ () (The transformaton from {0, 1} valued matrces to { 1, +1} means we have to work wth 2ɛ approxmatons rather than ɛ.) We choose = IP n. Then the numerator s (1 2ɛ)2 2n. It remans to upper bound µ (IP n ). µ (IP n ) = max x,y {0,1} 2n xt IP n y max 2n/2 IP n y y {0,1} 2n 6

by the Cauchy-Schwarz nequalty. Let us now evaluate IP n y 2 = y t IP t nip n y = 2 n y t y 2 2n. Puttng everythng together we have µ 2ɛ (IP n ) (1 2ɛ)2 n/2. Ths shows that R ɛ (IP n ) n/2 + log(1 2ɛ). References [1] T. S. Jayram, R. Kumar, D. Svakumar: Two applcatons of nformaton complexty. n Proceedngs of the 35th annual ACM symposum on Theory of computng, 2003, 673-682. [2] E. Kushlevtz, N. Nssan: Communcaton complexty. Cambrdge Unversty Press, 1997. [3] T. Lee, A. Shrabman: Lower bounds n communcaton complexty. Foundatons and Trends n Theoretcal Computer Scence 3(4): 263-398 (2009). 7