Theoretical Computer ScienCe 106 (1992) EIsevier. Îï the distributional complexity

Similar documents
Randomized Simultaneous Messages: Solution of a Problem of Yao in Communication Complexity

On the P versus NP intersected with co-np question in communication complexity


1 Sequences of events and their limits

Self-Testing Polynomial Functions Efficiently and over Rational Domains

Nondeterminism LECTURE Nondeterminism as a proof system. University of California, Los Angeles CS 289A Communication Complexity

Chain Independence and Common Information

Math 421, Homework #6 Solutions. (1) Let E R n Show that = (E c ) o, i.e. the complement of the closure is the interior of the complement.

Lecture 4: Proof of Shannon s theorem and an explicit code

Part V. 17 Introduction: What are measures and why measurable sets. Lebesgue Integration Theory

On non-antipodal binary completely regular codes

Mathematics 220 Workshop Cardinality. Some harder problems on cardinality.

Soft-decision Decoding of Chinese Remainder Codes

AN ARCSINE LAW FOR MARKOV CHAINS

Lemma 8: Suppose the N by N matrix A has the following block upper triangular form:

Solving a linear equation in a set of integers II

Quantum Communication Complexity

Lecture 16: Communication Complexity

Citation Osaka Journal of Mathematics. 43(2)

MAX -PLANCK -INSTITUT FUR INFORMATIK. On Multi-Party Communication Complexity of Random Functions. Technical Report No. MPII

THE S 1 -EQUIVARIANT COHOMOLOGY RINGS OF (n k, k) SPRINGER VARIETIES

Recitation 6. Randomization. 6.1 Announcements. RandomLab has been released, and is due Monday, October 2. It s worth 100 points.

Lecture 8: Probability

Note that in the example in Lecture 1, the state Home is recurrent (and even absorbing), but all other states are transient. f ii (n) f ii = n=1 < +

Discrete Applied Mathematics

Lecture 3: Lower Bounds for Bandit Algorithms

II - REAL ANALYSIS. This property gives us a way to extend the notion of content to finite unions of rectangles: we define

Slice rank of tensors and its applications

The integers. Chapter 3

MONOTONE SUBNETS IN PARTIALLY ORDERED SETS

CSE 20 DISCRETE MATH SPRING

Partitions and Covers

Geometric interpretation of signals: background

T i t l e o f t h e w o r k : L a M a r e a Y o k o h a m a. A r t i s t : M a r i a n o P e n s o t t i ( P l a y w r i g h t, D i r e c t o r )

Applications of Discrete Mathematics to the Analysis of Algorithms

Quantum Communication

Chapter 2. Linear Algebra. rather simple and learning them will eventually allow us to explain the strange results of

1/12/05: sec 3.1 and my article: How good is the Lebesgue measure?, Math. Intelligencer 11(2) (1989),

Lecture 3: Randomness in Computation

10.4 The Kruskal Katona theorem

Exercises 1 - Solutions

Tuples of Disjoint NP-Sets

Learning convex bodies is hard

Distribution of Environments in Formal Measures of Intelligence: Extended Version

STRONGLY EXTREME POINTS IN LPfa X)

New Imperfect Random Source with Applications to Coin-Flipping

Hardness of MST Construction

1 Randomized Computation

The Unbounded-Error Communication Complexity of Symmetric Functions

Linear Algebra in Computer Vision. Lecture2: Basic Linear Algebra & Probability. Vector. Vector Operations

18.5 Crossings and incidences

INDEPENDENCE THEORIES AND GENERALIZED ZERO-ONE LAWS

Linear Codes, Target Function Classes, and Network Computing Capacity

Ahlswede Khachatrian Theorems: Weighted, Infinite, and Hamming

CONVEX OPTIMIZATION OVER POSITIVE POLYNOMIALS AND FILTER DESIGN. Y. Genin, Y. Hachez, Yu. Nesterov, P. Van Dooren

CS Communication Complexity: Applications and New Directions

Learnability and the Doubling Dimension

Topological properties

hal , version 1-27 Mar 2014

18.10 Addendum: Arbitrary number of pigeons

Note. On the Upper Bound of the Size of the r-cover-free Families*

A Simple Left-to-Right Algorithm for Minimal Weight Signed Radix-r Representations

Chebyshev Polynomials, Approximate Degree, and Their Applications

Block Sensitivity of Minterm-Transitive Functions

arxiv:math/ v1 [math.lo] 5 Mar 2007

ASYMPTOTIC MAXIMUM PRINCIPLE

Continuum Probability and Sets of Measure Zero

Quick Sort Notes , Spring 2010

A diametric theorem for edges. R. Ahlswede and L.H. Khachatrian

On the number of occurrences of all short factors in almost all words

Disjointness conditions in free products of. distributive lattices: An application of Ramsay's theorem. Harry Lakser< 1)

P = 1 F m(p ) = IP = P I = f(i) = QI = IQ = 1 F m(p ) = Q, so we are done.

Containment restrictions

On Minimal Words With Given Subword Complexity

arxiv: v4 [cs.dm] 27 Oct 2016

Economics 204 Fall 2011 Problem Set 1 Suggested Solutions

Deformations of calibrated D-branes in flux generalized complex manifolds

A note on monotone real circuits

Notes 6 : First and second moment methods

VC-DENSITY FOR TREES

Automata and Languages

CS Foundations of Communication Complexity

On Randomized One-Round Communication Complexity. Ilan Kremer Noam Nisan Dana Ron. Hebrew University, Jerusalem 91904, Israel

ÇÙÐ Ò ½º ÅÙÐ ÔÐ ÔÓÐÝÐÓ Ö Ñ Ò Ú Ö Ð Ú Ö Ð ¾º Ä Ò Ö Ö Ù Ð Ý Ó ËÝÑ ÒÞ ÔÓÐÝÒÓÑ Ð º Ì ÛÓ¹ÐÓÓÔ ÙÒÖ Ö Ô Û Ö Ö ÖÝ Ñ ¹ ÝÓÒ ÑÙÐ ÔÐ ÔÓÐÝÐÓ Ö Ñ

Introduction. 1 Partially supported by a grant from The John Templeton Foundation

A Simple Left-to-Right Algorithm for Minimal Weight Signed Radix-r Representations

2 Plain Kolmogorov Complexity

c 2010 Society for Industrial and Applied Mathematics

Finite connectivity in infinite matroids

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

Bias and No Free Lunch in Formal Measures of Intelligence

Simultaneous Communication Protocols with Quantum and Classical Messages

DISTRIBUTIVE LATTICES ON GRAPH ORIENTATIONS

Lecture 5: February 21, 2017

MATH 802: ENUMERATIVE COMBINATORICS ASSIGNMENT 2

Great Expectations. Part I: On the Customizability of Generalized Expected Utility*

An Improved Quantum Fourier Transform Algorithm and Applications

B œ c " " ã B œ c 8 8. such that substituting these values for the B 3 's will make all the equations true

Lecture 2: Program Obfuscation - II April 1, 2009

Information Theory. Week 4 Compressing streams. Iain Murray,

Transcription:

Theoretical Computer ScienCe 106 (1992) 385-390 385 EIsevier Note Îï the distributional complexity of disjointness À.À. Razborov Steklov Mathematical Institute, Vavilova, 42, GSP-I, [17966, Moscow, Russia Communicated Üó M.S. Paterson Received October 1990 Revised October 1991 Abstract Razborov, À.À., Îï (1992) 385-390. the distributional complexity of disjointness, Theoretical Computer ScienCe 106 We prove that the distributional communication complexity of the predicate "disjointness" with respect to à very simple measure îï inputs is Ùï). 1. Introduction ÒÜå following concept of the G-error probabilistic communication complexity Ñ.(À) of à binary predicate À(õ, ó) was introduced Üó Óào [3]. Assume that two infinitely powerful computers evaluate the predicate À(õ, ó) in the situation when the first computer possesses õ and the second possesses ó (õ and ó are binary strings of length ï). ÒÜåó do this Üó interchanging messages between åàñü other. Both'Computers are allowed to flip à coin. At the end of the communication for åàñü õ and ó they must output the correct value of À(õ, ó) with probability at least 1-â. ÒÜå complexity is measured Üó the expected number of communications in the worst case. For more details see [3, 1]. In [4], Óào suggested àï approach to estimating Ñ.(À) from below and gave àï application of this approach. It is based èðîï the notion of the G-error distributional communication complexity D.(A) of à binary predicate À(õ, ó). This notion, in turn, was generalized in [1] to the concept of the G-error distributional communication 0304-3975/92/$05.00 (Q) 1992-EIsevier Science Publishers Â.Ó. AII rights reserved

-- Ç86 À.À. Razborov complexity De(A I Jl) under àï arbitrary probabilistic measure Jl îï inputs (De(A) is just De(A I Jl) with uniform Jl). This concept is somewhat dual to Ñã(À): now the computers run à deterministic protocol and are required to output the correct value of À(õ, ó) everywhere except for at most e-fraction (with respect to the measure Jl) of inputs. It was proved in [4] for uniform Jl and generalized in [1] to arbitrary Jl that Ce(A);::::tD2e(AiJl) for any A,Jl and å>o. Several authors studied these complexity measures for the predicate "disjointness". Let DISn denote this predicate (we will recall its definition below). Babai et al. [1] proved that De(DISnIJl);::::.Q(nl/2), where Jl is some measure on inputs and å>o is sufficiently small. This implies Ce(DISn);::::.Q(nl/2) for any e<t. ÒÜå measure Jl in [1] is à product measure that is the product of à measure on columns and à measure on rows. In comparison with the lower bound it was also proved in [1] that De(DISnIJl)::;::;O(nl/21ogn) for any product measure Jl and arbitrary å>o. Then Kalyanasundaram and Schnitger [2] established the best possible lower bound Ce(DISn);::::.Q(n) (å< 1/2) for the e-error probabilistic communication complexity of "disjointness". Probably alllower bounds for Ñã(À) known prior to the paper [2] were actually lower bounds for the distributional complexity De(A I Jl) with some suitable measure Jl. But the proof in [2] involves complicated arguments related to the Kolmogorov complexity and this results in the fact that the measure Jl implicitly "nieant" in the proof depends on the protocol given Üó "the adversary". ÒÜå aim of this note is to show that the "random coupling" arguments of Kalyanasundaram and Schnitger can Üå carried over to yield the lower bound De(DISIi I Jl ) ;:::: Ï( ï) for à very simple measure Jl described below (this does not contradict the result from [ 1] since our Jl is not à product measure ). ÒÜå proof involves only classical probabilistic arguments and does not appeal to the Kolmogorov complexity. 2. ÒÜå result We will identify throughout binary predicates and their characteristic 0-1 matrices. Given à predicate À(õ, ó) (ÕÅÕ, ÓÅ Ó), the e-error distributional complexity De(A I Jl) under à probabilistic measure Jl îï inputs (i.e., on Õ õ Ó) is the minimal possible length of à deterministic communication protocol which, given the random input (õ, ó) according to the measure Jl, outputs àõó with probability at least 1-å [4,1]. Fix the notation DISn for the so-called disjointness matrix DISn over Õ := Ó:= f!jj([n]) given Üó (DISn)xy:= 1 iff õïó=ô. Let (Õo,óo) [(Õl, Ól )] Üå the random input according to the uniform distribution on {(õ, y)llxl =Iyl =Ln/4 J, Ixnyl =ô} [ {(õ, y)llxl =Iyl =Ln/4 J, Ixnyl=l}, respectively}. Let (õ,ó) Üå taken with probability i as (Õo,óo) and with probability t as (Õl,Ól). Denote Üó Jl the measure corresponding to (õ,ó). ÒÜå main result of this note is the following theorem.

Îï the distributional complexity î! disjointness 387 Theorem. D.(DISn l.è);;;:q(ï)for àïó sufficiently small Â>O. Proof. Note that (õ,,' ó,,) (VE{O, 1} ) isjust the input (õ, ó) under the condition 'õïó! =V. Because ofp[lxnyl=v];;;:q(l) the theorem follows from the following statement (cf. [4,1]). Main Lemma. For àïó Õ, Ys;;f!/([n]), Ð[(Õl'Ól)ÅÕ Õ Ó] ;;;:à(ð[(õî,óî)åõ Õ Ó])-2-Ùï). Proofofmain lemma. We òàó assume ï=4ò-1 and, therefore, Ixl=lyl=m. First we need à somewhat exotic way of generating random inputs (Õo,Óo) and (Õl'Ól). Namely, let t :=(Zx, Zy, { i} ) üå the random partition of [ï] into three sets of cardinalities 2ò-1, 2ò-1 and 1, respectively. Let õ üå the random member of[zxu{i} ]ò,ó Üå the random member of [Zyu{ i} ]ò (õ andy are assumed to Üå independent). Let (õo, óo) üå this (õ, ó) under the condition iôx, iôy and (Õl,Ól) Üå (õ,ó) under the condition iex, iey. Note that ur construction is invariant under the action ofthe symmetric group Sn; therefore, we obtain in this way the random inputs corresponding to the uniform distributions on {(õ, y)llxl=lyl=m, Ixnyl=0} and {(õ,ó) Ilxl=lyl=m, Ixnyl= l},i.e.,exactly(xo,yo) and (õ 1, ó 1) used in the definition of.è (itis also easy to see that (õ, ó) coincides with ur main distribution but we will not need this fact in what follows). Given à partition t = (zx, Zy, { i} ), set Px(t) :=Ð[ÕÅÕ I (Zx, Zy, { i} )=t], py(t) :=Ð[ÓÅ Ó! (Zx, Zy, { i} )=t], Px,o(t) :=P[xoEXI(zx, Zy, {i})=t], Ðõ, 1 (t):=p[xl ÅÕ I (Zx, Zy, { i} )=t], Ðó, o(t):=p[yoe Ó! (Zx, Zy, { i} )=t], Py.l(t):=P[ylEYI(zx,zy, {i})=t]. Then Ð[(õo,óo)ÅÕÕ Ð[(Õl'Ól)ÅÕ Y]=E[px.o(t).py.o(t)], Õ Y]=E[px.l(t).Py.l(t)]. We now collect some easy facts about these random variabies. Fact 1. Px(t)=!(Px.o(t)+PX.l(t)), Py(t)=!(py,o(t)+Py.l(t)).

388 À.À. Razborov Proof. The result follows from the observation P[iEXI(zx,Zy, {i})=t]= P[iEYI(zx,Zy,{i})=t]=!. D Fact 2. Px(zx, Zy, { i} ) and Ðó, o(zx, Zy, { i} ) depend only îï Ðõ, o(zx, Zy, { i} ) depend only îï zx. Zy. Py(zx, Zy, { i} ) and Set å := 0.01. Let us say that t is x-bad if Px,1(t)<1Px,o(t)-2-en (1) and t is y-bad if Ðó, 1 (t) <ipy, o(t)- 2 -åï. t is bad if it is either x-bad or y-bad. Claim 3. For àïó Zx,ZyE[n]2m-l,P[(zx,zy,{i}) is x-badlzy=zy]<! and P[(zx, Zy, {i}) is y-bad I zx=zx]<t. Proof. Âó symmetry, it is sufficient to prove the first inequality. Âó Fact 2, having fixed Zy forces Ðõ( Zx, Zy, { i} ) to üå constant. Denote Ðõ( Zx, Zy, { i} ) Üó Ðõ. If Ðõ < 2 -.ï then, Üó (1) and Ðõ, o(t)~2px(t) (see Fact 1), P[(zx, Zy, {i} ) is x-bad I Zy=Zy] =0 and we are done. So, assume px~2-eï. (2) Denote Xn[co-zy]m Üó S. Then Px=ISI/(2:);px,1(Zx,Zy, {i})=2pxp[ies]; Px,o(zx, Zy, {i})=2pxp[iis], where S is the random member ofs. So, if(zx, zy,{i}) is x-bad then, Üó (1), P[iES] ~1P[iis], (3) i.e., P[iES]~i. Îî the other hand, s=(s~,s2'...'s2m), where Sl,S2,...,S2m are the characteristic functions of events i1es, i2es,..., i2mes ({i1, i2,..., i2m}=co-zy). Assume, contrary to the statement ofthe claim, that Ð [(zx, Zy, {i} ) is x-bad I Zy=Zy] ~! and, hence, (3) holds for at least 2ò/5 values of ieco -Zy. Then, counting the entropy, we get m(2-4e-o(1))~h(s) (Üó (2))~ L H(s;)~8m/5+2m/5.H(1/4)~1.93m, i= 1 2ò à contradiction. D bad]. Let us denote Üó Xx(t) [Xy(t), X(t)] the indicator of the event "t is x-bad" [y-bad, Claim 4. E[px,o(t)py,o(t)X(t)] ~!E[px,o(t)py,o(t)].

~ Îï the distributional cornplexity î! disjointness Ç89 ProoC. Because X(t) :::;;Xx(t) + Õó( t), itis sufficient to prove that Å[ðõ, o(t) Ðó, o(t)xx(t)] :::;; ~E[px,o(t)py,o(t)]. Let us fix Zy and prove that Å[ðõ, o(zx, Zy, {;} )Ðó,o (Zx, Zy, { ;} ) Xx(Zx, Zy, { ;} ) I Zy=Zy] :::;;~E[px,o(Zx, Zy, {;})Py,o(Zx, Zy, {;})Izy=zy]. Note that Ðó, o(zx, Zy, { ;} ) and Px(Zx, Zy, { ;} ) are constant under the condition Zy = Zy (see Fact 2). Denote them Üó Ðó, î and Ðõ. It is also clear that Å[ðõ, o(zx, Zy, { ;} ) I Zy=Zy] = Ðõ since this expectation is just Ð[ÕoÅÕ I Zy =Zy] and Õo under the only condition Zy=Zy takes all values Crom [co-zy]m with the same probability ( 2:) -1, i.e., coincides with õ under the same condition. ThereCore, Å[ðõ, o(zx, Zy, { ;} )Ðó, o(zx, Zy, { ;} )Xx(Zx, Zy, { ;}) I Zy =Zy] = Ðó, oå [Ðõ, o(zx, Zy, { ;} ) Xx(Zx, Zy, { ;} ) I Zy = Zy] :::;; 2ðó, oðõå [Õõ( Zx, Zy, { ;} ) I Zy = Zy] (Üó Fact 1) :::;; ~ Ðó, oðõ (Üó Claim 3) =ipy,oe[px,o(zx, Zy, {;} ) I Zy=Zy] =ie[px,o(zx,zy,{;})py,o(zx,zy,{;})lzy=zy]. (J ProoC îñ main lemma (conclusion). Now the prooc îñ the main lemma is completed Üó the easy computation Ð[(Õ1,Ó1)ÅÕ õ Y]=E[px,l(t).Py,l(t)]~E[Px,l(t).Py,l(t).(l-X(t)] ~ Å [(ipx, o(t)- 2-åï).(ipy. o(t)- 2-åï).(1- X(t)] (Üó(l)) ~.Q(E[px, o(t).ðó, o(t).(1- X(t)])- 2 -Î(ï) ~.Q(E[px,o(t).py,o(t)])-2-n(n) (Üó Claim 4) =.Q(P[(Xo,yo)EX õ Ó])-2-Î(ï). (J Acknowledgment I àò thankcul to Laci Babai and Nathan Linial Cor valuable remarks.

390 À.À. Razborov References [1] L. Babai, Ð. Frankl and J. Simon, Complexity classes in communication complexity theory, in: Proc. 27th IEEE FOCS (1986) 337-347. [2] Â. Kalyanasundaram and G. Schnitger, The probabilistic communication complexity of set intersection, in: Proc. 2nd Àïï. Ñîï! îï Structure jn Complexjty Theory (1987) 41-49. [3] À.Ñ. Óàî, Some complexity questions related to distributed computing, in: Proc. llth ÀÑÌ STOC (1979) 209-213. [4] À.Ñ. Óào, Lower bounds Üó probabilistic arguments, in: Proc. 24th IEEE FOCS (1983) 420-428. * 1 (;,;"..ò 1é -!(",1'(,'"Ç;:).i!'l ii'i