Lecture 20: Lower Bounds for Inner Product & Indexing

Similar documents
Communication Complexity. The dialogues of Alice and Bob...

Partitions and Covers

15-251: Great Theoretical Ideas In Computer Science Recitation 9 : Randomized Algorithms and Communication Complexity Solutions

Communication Complexity

CS Introduction to Complexity Theory. Lecture #11: Dec 8th, 2015

Lecture Lecture 9 October 1, 2015

Lecture 21: Quantum communication complexity

Lecture 16 Oct 21, 2014

Lecture 11: Quantum Information III - Source Coding

Randomness. What next?

Course Introduction. 1.1 Overview. Lecture 1. Scribe: Tore Frederiksen

Matrix Rank in Communication Complexity

Nondeterminism LECTURE Nondeterminism as a proof system. University of California, Los Angeles CS 289A Communication Complexity

CS Foundations of Communication Complexity

CS Communication Complexity: Applications and New Directions

The Communication Complexity of Correlation. Prahladh Harsha Rahul Jain David McAllester Jaikumar Radhakrishnan

25.2 Last Time: Matrix Multiplication in Streaming Model

An Information Complexity Approach to the Inner Product Problem

Asymmetric Communication Complexity and Data Structure Lower Bounds

14. Direct Sum (Part 1) - Introduction

Lecture 8: Channel and source-channel coding theorems; BEC & linear codes. 1 Intuitive justification for upper bound on channel capacity

Lecture 11: Polar codes construction

Lecture 11: Key Agreement

CSCI-B609: A Theorist s Toolkit, Fall 2016 Oct 4. Theorem 1. A non-zero, univariate polynomial with degree d has at most d roots.

Lecture 18: Quantum Information Theory and Holevo s Bound

Lecture 16: Communication Complexity

CS Foundations of Communication Complexity

LECTURE 13. Last time: Lecture outline

Approximation norms and duality for communication complexity lower bounds

Communication vs information complexity, relative discrepancy and other lower bounds

An exponential separation between quantum and classical one-way communication complexity

Lecture 4: Proof of Shannon s theorem and an explicit code

On Function Computation with Privacy and Secrecy Constraints

1 Review of The Learning Setting

Lecture 4: Codes based on Concatenation

Information Complexity and Applications. Mark Braverman Princeton University and IAS FoCM 17 July 17, 2017

Communication is bounded by root of rank

Tutorial on Quantum Computing. Vwani P. Roychowdhury. Lecture 1: Introduction

Lecture 15 April 9, 2007

Lecture 6 Sept. 14, 2015

6.895 PCP and Hardness of Approximation MIT, Fall Lecture 3: Coding Theory

Compute the Fourier transform on the first register to get x {0,1} n x 0.

ECE598: Information-theoretic methods in high-dimensional statistics Spring 2016

CODING AND CRYPTOLOGY III CRYPTOLOGY EXERCISES. The questions with a * are extension questions, and will not be included in the assignment.

Lecture 5: February 21, 2017

Communication Complexity

Direct product theorem for discrepancy

1 Basic Information Theory

Lecture 6: Lattice Trapdoor Constructions

1 Maintaining a Dictionary

Multiparty Computation

MATH 112 QUADRATIC AND BILINEAR FORMS NOVEMBER 24, Bilinear forms

Communication Complexity 16:198:671 2/15/2010. Lecture 6. P (x) log

Lecture: Analysis of Algorithms (CS )

Lecture 6: Expander Codes

Math Linear Algebra II. 1. Inner Products and Norms

1 Randomized Computation

Hardness of MST Construction

Lecture 8: Shannon s Noise Models

Direct product theorem for discrepancy

SIGNATURE SCHEMES & CRYPTOGRAPHIC HASH FUNCTIONS. CIS 400/628 Spring 2005 Introduction to Cryptography

Quantum sampling of mixed states

14 Diffie-Hellman Key Agreement

Theoretical Cryptography, Lecture 10

Lecture 5, CPA Secure Encryption from PRFs

Fourier analysis of boolean functions in quantum computation

Lecture 6 I. CHANNEL CODING. X n (m) P Y X

LECTURE 2. Convexity and related notions. Last time: mutual information: definitions and properties. Lecture outline

b = 10 a, is the logarithm of b to the base 10. Changing the base to e we obtain natural logarithms, so a = ln b means that b = e a.

Lecture 5. 1 Review (Pairwise Independence and Derandomization)

6.896 Quantum Complexity Theory November 4th, Lecture 18

Communication with Contextual Uncertainty

Lecture 22. We first consider some constructions of standard commitment schemes. 2.1 Constructions Based on One-Way (Trapdoor) Permutations

Lecture 6. Today we shall use graph entropy to improve the obvious lower bound on good hash functions.

Lecture 10 Oct. 3, 2017

Quantum Communication

Message Authentication

EE/Stat 376B Handout #5 Network Information Theory October, 14, Homework Set #2 Solutions

Introduction to Cryptography Lecture 13

Chapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University

MATH41011/MATH61011: FOURIER SERIES AND LEBESGUE INTEGRATION. Extra Reading Material for Level 4 and Level 6

Lecture 10: NMAC, HMAC and Number Theory

Separating Deterministic from Nondeterministic NOF Multiparty Communication Complexity

Exponential Separation of Quantum Communication and Classical Information

Interactive Channel Capacity

LOWER BOUNDS FOR QUANTUM COMMUNICATION COMPLEXITY

Information Complexity vs. Communication Complexity: Hidden Layers Game

Lecture 4: Constructing the Integers, Rationals and Reals

Lecture Notes, Week 6

Lecture 26: Arthur-Merlin Games

1 Review of Winnow Algorithm

Lecture 12: November 6, 2017

Lecture 8: Linearity and Assignment Testing

ROUNDS IN COMMUNICATION COMPLEXITY REVISITED

Direct product theorem for discrepancy

Notes 3: Stochastic channels and noisy coding theorem bound. 1 Model of information communication and noisy channel

The Hardness of Being Private

Introduction to Machine Learning (67577) Lecture 3

Assignment 1: From the Definition of Convexity to Helley Theorem

Lecture 15: Expanders

Transcription:

15-859: Information Theory and Applications in TCS CMU: Spring 201 Lecture 20: Lower Bounds for Inner Product & Indexing April 9, 201 Lecturer: Venkatesan Guruswami Scribe: Albert Gu 1 Recap Last class Randomized Communication Complexity Distributional CC : D µ δ (f) is the best communication complexity of a deterministic protocol Π such that Pr µ [Π(x, y) f(x, y)] δ. Lemma: R pub δ (f) = max µ D µ δ (f) can be used to lower bound R(f) by choosing an adverse distribution µ. Today Lower bounds on Distributional CC Discrepancy Indexing problem via Information Theory 2 Discrepancy technique We would like to develop a method to lower bound D µ δ, which in turn will lower bound R(f). Every deterministic protocol induces a partition of X Y into rectangles; previously, for zeroerror protocols, these rectangles had to be monochromatic, but now we allow some to not be monochromatic. The discrepancy technique aims to show that, for a specific function f, every large rectangle has nearly equal numbers of 0s and 1s. This forces an accurate protocol to use only small rectangles and hence require many rectangles. Definition 1 (Discrepancy). Let f : X Y {0, 1}, R = S T : S X, T Y, and µ be a distribution on X Y. Denote Disc µ (R, f) = = Pr (x,y) R [f(x, y) = 0 (x, y) R] Pr [f(x, y) = 1 (x, y) R] µ ( 1) f(x,y) µ(x, y) 1

and Disc µ (f) = (Note: If R is monochromatic, Disc µ (R, f) = µ(r).) max Disc µ(r, f) R X Y Note that large discrepancy (monochromatic and large rectangle) ( is good for a protocol. ) 1 following is a generalization of the deterministic bound D(f) log 2 : Proposition 2 (Discrepancy lower bound). D µ 1 2 γ(f) log 2 ( 2γ Disc µ(f) max Rmonochr µ(r) ) The Proof. Let Π be a protocol using c bits of communication with error probability at most 1 2 γ. Since it is deterministic, the matrix is split into at most 2 c rectangles. By the maximum error allowed from this protocol, we have Pr [Π(x, y) = f(x, y)] Pr [Π(x, y) f(x, y)] 2γ We can bound the LHS by breaking it into rectangles according to the protocol and noting that Π is constant on each rectangle: Pr [Π(x, y) = f(x, y)] Pr [Π(x, y) f(x, y)] = Pr [Π(x, y) = f(x, y) (x, y) R l ] µ µ µ This implies 2 c Disc µ (f) 2γ = c log 2 ( R l protocol Pr [Π(x, y) f(x, y) (x, y) R l ] µ Pr [f(x, y) = 0 (x, y) R l ] [f(x, y) = 1 (x, y) R l ] µ R l = R l Disc µ (R l, f) 2 c Disc µ (f) 2γ Disc µ(f) ), as desired. 2.1 Dot product function We will now apply this technique to bound the randomized CC of the dot product function, defined as IP (x, y) = x y = x i y i (mod 2) In the deterministic case, we showed in a previous lecture that n + 1 is the best we could do. Theorem. R 1 (IP ) Ω(n) = n 2 O(1) 2

It suffices to show that D µ 1 (IP ) n 2 O(1) for some distribution µ. This has two parts: we need to come up with a clever µ, and then need to bound it. Since dot product is pretty evenly distributed for random inputs, we take µ to be uniform. Goal: Prove Disc uniform (IP ) 1 2 n/2 (Note that this implies the claimed bound by the above Proposition). Proof. Let R = S T be any rectangle. Then Disc µ (R, IP ) = x S,y T ( 1) x y 1 Let H n {1, 1} 2n 2 n be the matrix indexed by X and Y where the (x, y)th entry is ( 1) x y. First we show the following fact. Exercise: H n is an orthogonal matrix (H t nh n = 2 n I). Now we can bound Disc µ (R, IP ): 2 2n Disc µ (R, IP ) = 1 2 2n 1 S t H n 1 T = 1 2 2n (1 S t ) (H n 1 T ) 1 2 2n 1 S H n 1 T S (Hn 2 2n 1 T ) (H n 1 T ) S 2 2n 1 t T Ht nh n 1 T S 2 2 2n n T 1 2 2 2n n 2 n 2 n = 1 2 n/2 (Note: It is possible to improve the bound for R(IP ) to n O(1), which appears on Problem Set 4) In summary, we have shown that R(EQ) = θ(log n) and R(IP ) = θ(n). In upcoming lectures, we will tackle R(DISJ), which is in some sense the poster child of this whole field.

Indexing Problem Alice and Bob are again communicating, but the setup is slightly asymmetrical this time. As before, Alice has a string x {0, 1} n, but now Bob has an index i {1, 2,, n}, and the goal is for Bob to learn x i. There is a trivial log n protocol by just sending the index. Now, suppose we only allow Alice to send a single message so that Bob can figure out x i. Can we do better than the trivial n bit solution?.1 Deterministic Suppose Alice and Bob use a deterministic protocol and Alice sends less than n bits. Then there exists a b such that Alice sends the same message for a, b and Bob cannot distinguish if Alice sent A or B. Let j be such that a j b j, then the protocol is wrong on either (a, j) or (b, j)..2 Randomized The above proof does not give us enough for a randomized lower bound: we need Bob to be wrong on a lot of inputs. However, it turns out that even a randomized protocol requires Ω(n) bits to be sent, which can be shown in several ways. Exercise: Come up with a µ on {0, 1} n {1,, n} such that D µ 1/ (Index) Ω(n). Hint: If a and b in the deterministic proof differ in only 1 bit, Bob has low chance of error. We would like them to differ in more. Try finding a distribution on X supported on a code of distance n/, and also supported on 2 Ω(n) elements. In contrast to the coding theory proof hinted at above, we will present an information theoretical proof of this fact. Proof. We will bound the distributional complexity. We take the distribution X = X 1 X 2 X n uniform on {0, 1} n and i uniform on {1,, n}. Let Π be a deterministic protocol with error at most 1. Alice will send M = M(x), also a random variable. We can bound CC(Π) log(supp(m)) H(M) = I(M; X) = I(X 1 X 2 X n ; M) I(X i ; M) The goal is now to show that M has a lot of information, since Bob can tell a lot about X from M. We would like to show that each I(X i ; M) is about a constant, which makes sense since Bob can figure out any bit with high probability. Continuing the chain of inequalities, CC(Π) = I(X i ; M) = H(X i ) H(X i M) = n H(X i M) 4

For notation, let Pe m,i be the probability of error given that Alice sent m and Bob has i. By the error guarantee of the protocol, we have E m,i [P i e] 1 By Fano s Inequality, h(pe m,i ) H(X i M = m). Therefore Finally, by the concavity of h, this gives i E m,i [h(p i e)] = E i [E m [h(p m,i e )]] E i [E m [H(X i M = m)]] H(X i M) E[h(P m,i e E i [H(X i M)] i = H(X i M) n )]n h ( E[P m,i ] ) n h e ( ) 1 n Wrapping it all up, we have CC(Π) n i H(X i M) n h ( ) 1 n Ω(n). 5