Tutorial: Locally decodable codes. UT Austin

Similar documents
Three Query Locally Decodable Codes with Higher Correctness Require Exponential Length

A Quadratic Lower Bound for Three-Query Linear Locally Decodable Codes over Any Field

Local correctability of expander codes

Locally Decodable Codes

Locally Decodable Codes

Lecture Introduction. 2 Formal Definition. CS CTT Current Topics in Theoretical CS Oct 30, 2012

Improved Lower Bounds for Locally Decodable Codes and Private Information Retrieval

Corruption and Recovery-Efficient Locally Decodable Codes

A list-decodable code with local encoding and decoding

An approach from classical information theory to lower bounds for smooth codes

High-Rate Codes with Sublinear-Time Decoding

Error Detection and Correction: Hamming Code; Reed-Muller Code

2-LOCAL RANDOM REDUCTIONS TO 3-VALUED FUNCTIONS

ALL codes discussed in this paper are linear. We study. Locally Testable Cyclic Codes. László Babai, Amir Shpilka, and Daniel Štefankovič

Private Locally Decodable Codes

Private Locally Decodable Codes

High-rate codes with sublinear-time decoding

Lecture 8 (Notes) 1. The book Computational Complexity: A Modern Approach by Sanjeev Arora and Boaz Barak;

Two Query PCP with Sub-Constant Error

1 Randomized Computation

Exponential Lower Bound for 2-Query Locally Decodable Codes via a Quantum Argument

Locality in Coding Theory

Constant Query Locally Decodable Codes against a Computationally Bounded Adversary

LOCALLY DECODABLE QUANTUM CODES

6.895 PCP and Hardness of Approximation MIT, Fall Lecture 3: Coding Theory

CS294: Pseudorandomness and Combinatorial Constructions September 13, Notes for Lecture 5

CS151 Complexity Theory. Lecture 9 May 1, 2017

Low Rate Is Insufficient for Local Testability

Locally Testable Codes and PCPs of Almost-Linear Length

Lecture 6. k+1 n, wherein n =, is defined for a given

Efficient Probabilistically Checkable Debates

Noisy Interpolating Sets for Low Degree Polynomials

Breaking the O(n 1/(2k 1) ) Barrier for Information-Theoretic Private Information Retrieval

Tolerant Versus Intolerant Testing for Boolean Properties

Testing Low-Degree Polynomials over GF (2)

Tolerant Versus Intolerant Testing for Boolean Properties

Almost transparent short proofs for NP R

Lecture 4: Codes based on Concatenation

Error-Correcting Codes:

HARDNESS AMPLIFICATION VIA SPACE-EFFICIENT DIRECT PRODUCTS

Example: sending one bit of information across noisy channel. Effects of the noise: flip the bit with probability p.

SIGACT News Complexity Theory Column 25

Lecture 3: Error Correcting Codes

arxiv: v1 [cs.cr] 16 Dec 2014

Sparse Random Linear Codes are Locally Decodable and Testable

Noisy Interpolating Sets for Low Degree Polynomials

Finite fields, randomness and complexity. Swastik Kopparty Rutgers University

LOCALLY DECODABLE CODES WITH TWO QUERIES AND POLYNOMIAL IDENTITY TESTING FOR DEPTH 3 CIRCUITS

Lecture 21: P vs BPP 2

Proposal Summary Intellectual Merit: Broader Impact: Keywords:

Notes for the Hong Kong Lectures on Algorithmic Coding Theory. Luca Trevisan. January 7, 2007

Report on PIR with Low Storage Overhead

Computability and Complexity Theory: An Introduction

2. Polynomials. 19 points. 3/3/3/3/3/4 Clearly indicate your correctly formatted answer: this is what is to be graded. No need to justify!

Quantum Property Testing

Lecture 12: November 6, 2017

Basic Probabilistic Checking 3

Quantum Symmetrically-Private Information Retrieval

Near-Optimal Secret Sharing and Error Correcting Codes in AC 0

Lecture 10: Hardness of approximating clique, FGLSS graph

Error Correction Review

PCP Theorem and Hardness of Approximation

Lecture 19 : Reed-Muller, Concatenation Codes & Decoding problem

Highly resilient correctors for polynomials

The Tensor Product of Two Codes is Not Necessarily Robustly Testable

Local list-decoding and testing of random linear codes from high-error

List and local error-correction

Pseudorandomness and Combinatorial Constructions

Relaxed Locally Correctable Codes in Computationally Bounded Channels

Quantum Communication Complexity

Lecture 6: Expander Codes

Locally Decodable Codes and Private Information Retrieval Schemes. Sergey Yekhanin

Private Information Retrieval from Transversal Designs

Quantum Property Testing

Lecture 19: Interactive Proofs and the PCP Theorem

Notes for Lecture 14 v0.9

arxiv: v1 [cs.it] 18 Jun 2011

Quantum walk algorithms

Shannon s noisy-channel theorem

Arnab Bhattacharyya Zeev Dvir Subhangi Saraf Amir Shpilka

ELEC 519A Selected Topics in Digital Communications: Information Theory. Hamming Codes and Bounds on Codes

Error Correcting Codes: Combinatorics, Algorithms and Applications Spring Homework Due Monday March 23, 2009 in class

Incidence geometry and applica1ons. Shubhangi Saraf Rutgers University

Reed-Muller testing and approximating small set expansion & hypergraph coloring

Enumeration: logical and algebraic approach

List Decoding of Reed Solomon Codes

Advanced Algorithms (XIII) Yijia Chen Fudan University

A Self-Tester for Linear Functions over the Integers with an Elementary Proof of Correctness

Lecture 15 - Zero Knowledge Proofs

Algebraic Geometry Codes. Shelly Manber. Linear Codes. Algebraic Geometry Codes. Example: Hermitian. Shelly Manber. Codes. Decoding.

Tool: Linearity Testing and the Walsh-Hadamard Code

Notes for Lecture 18

Last time, we described a pseudorandom generator that stretched its truly random input by one. If f is ( 1 2

The power and weakness of randomness (when you are short on time) Avi Wigderson Institute for Advanced Study

EE 229B ERROR CONTROL CODING Spring 2005

Pseudorandomness and combinatorial constructions

Algorithms for sparse analysis Lecture I: Background on sparse approximation

Errors, Eavesdroppers, and Enormous Matrices

Transcription:

Tutorial: Locally decodable codes Anna Gál UT Austin

Locally decodable codes Error correcting codes with extra property: Recover (any) one message bit, by reading only a small number of codeword bits. Motivation: large data sets sublinear decoding time Applications: Group Testing? :-) Private Information Retrieval Probabilistically Checkable Proofs Data Structures Polynomial Identity Testing Matrix Rigidity and circuit complexity

Traditional error correcting codes: C : {0, 1} k {0, 1} n x {0, 1} k can be uniquely reconstructed from any y {0, 1} n within Hamming distance δn from C(x). Decoding involves reading all of y. What if we only need to recover a small part of x?

Locally decodable codes: no need to read all encoded data to recover individual parts Examples: Phone book only want one phone number Hospital records only want record of one patient Don t want to read encoding of the whole phonebook or all records of the entire hospital

First attempt Encode each phone number (each patient record) separately. This would work if the errors follow a nice pattern: no more than δ fraction of every record BUT: δ fraction of all of the encoded data can be corrupted Some records can get completely destroyed this way: δ = 1/10, 10000 patients, 1000 can be lost.

Examples: Phone book only want one phone number Hospital records only want record of one patient Don t want to read encoding of the whole phonebook or all records of the entire hospital But want to be able to get record of ANY one of the patients (ANY phone number)

Ideally: amount of encoded data to read proportional to how many records I want to get. Or, at least: to get ONE record, read only a sublinear part BUT: δ fraction of all of the encoded data can be corrupted Can we do this?

Decoder must be probabilistic. Sublinear part of the data - much less than δn bits. If the decoder is deterministic, possibly all the bits it reads are corrupted

Examples of some trivial probabilistic decoders. We try to recover the bit x 1. 1. Flip a coin, return the result (don t read anything) Correct with probability 1/2 2. Suppose the code has many copies of x 1 in the codeword (repetition code for bit x 1 ) Randomly pick a position. Correct with probability 1 δ.

(q, δ, ɛ)-ldc decoder reads at most q bits tolerates up to δ fraction of errors in codeword correct with probability at least 1/2 + ɛ Better than random guessing if ɛ > 0

Locally decodable codes [Katz, Trevisan, 2000] C : {0, 1} k {0, 1} n is a (q, δ, ɛ)-ldc if probabilistic A s.t. 1. i [k], x {0, 1} k, y {0, 1} n s.t.d(c(x), y) δn P r[a(y, i) = x i ] 2 1 + ɛ 2. A reads q bits of y Advantage over random guessing ɛ Correctness 1 2 + ɛ For codes over Σ, 1 Σ + ɛ.

Efficient constructions: Linear n = O(k) length, sublinear q = k β queries (based on ideas from [Beaver, Feigenbaum 1990]). [Babai, Fortnow, Levin, Szegedy 1991]: nearly linear length n = k 1+β, q = polylog(k) No polynomial length codes known for constant number of queries.

1-query Locally Decodable Code not possible [Katz, Trevisan 2000] For any δ, ɛ > 0 there exist k δ,ɛ s.t. for k > k δ,ɛ (1, δ, ɛ)-ldc does not exist.

Hadamard code Based on ideas in [Blum, Luby, Rubinfeld 1990]: Hadamard code is (2, δ, 1/2 2δ) - LDC for any δ Correctness 1 2δ, nontrivial when δ < 1/4 n = 2 k Columns of generator matrix: all possible vectors of length k.

Hadamard code x 1,..., x k C(x) 1,..., C(x) n Each codeword bit C(x) j is some linear combination: C(x) j = a 1,j x 1 + a 2,j x 2 +... + a k,j x k 2-query decoder to find x i : Pick random j [n], read y j and y j s.t. C(x) j = C(x) j x i Correct with prob. at least 1 2δ

Lower bounds for 2-query codes Goldreich, Karloff, Schulman, Trevisan 02 For 2-query linear LDCs n 2 Ω(k) holds for any field [Dvir, Shpilka] Kerenidis, de Wolf 03 For 2-query binary LDCs n 2 Ω(k) Still open: nontrivial lower bounds for 2-query nonlinear codes over alphabets of size Ω( k).

Katz, Trevisan 00 n Ω(kq 1) q Kerenidis, de Wolf 03 n Ω((k/ log k) q+1 q 1) Woodruff 07 n Ω(k q+1 q 1)/ log k General lower bounds All below quadratic for q 3.

3-query linear codes Woodruff 2010 n Ω(k 2 ) for 3-query linear codes over any field. OPEN: 3-query linear locally decodable codes with n = O(k 2 ).

Subexponential size 3-query LINEAR codes Yekhanin 07 assuming infinitely many Mersenne primes Efremenko 08 unconditional n = 2 2O( log k log log k) k (log k)c 2 kα Correctness 1 3δ for nonbinary codes. Woodruff 08 Correctness 1 3δ η for binary codes.

Gal, Mills 2011 Tradeoffs between length and correctness 3-query codes require exponential size for higher correctness Correctness 1 3δ + 6δ 2 4δ 3 +γ requires size n 2 Ω(γ k) Holds for: LINEAR codes over any finite field BINARY NONLINEAR codes

Recovery graphs and matchings Hadamard code For each i [k] the possible pairs we query for x i form a matching. Pick random j [n], read y j and y j = y j + x i C(x) j + C(x) j = x i Given j, the corresponding j is unique (for fixed i) we get a matching.

Graph G = (V, E): vertices codeword positions {1,..., n} edges pairs of positions {j 1, j 2 } we query with probability > 0 Graph G of the Hadamard code: union of k matchings, each of size n/2

Linear codes: vector a j for each position C(x) j of the codeword: C(x) j = a j x. a j : j-th column of the generator matrix A: C(x) = A x. Binary codes: a j {0, 1} k. e i {0, 1} i-th unit vector (1 in position i, 0 everywhere else.) e i = a j1 + a j2 x i = C(x) j1 + C(x) j2 for every x. edges of the Boolean hypercube Hadamard code uses such pairs.

Combinatorial Lemma[GKST] Let M i be a set of disjoint pairs (j 1, j 2 ) such that e i = a j1 + a j2. Then n 2 2αk where α = ki=1 M i kn. [Katz, Trevisan 2000] The graph of every good 2-query locally decodable code must contain a large matching for EVERY i [k].

Extend to q-query codes q-tuples of positions hypergraphs or set systems. A q-tuple is a q-edge for x i if queried with prob. > 0 gives nontrivial advantage over random guessing.

Reed Muller codes r (prime power) alphabet size m number of variables of the polynomials d degree (d < r 1) k = ( ) m+d d message length = # of possible monomials message x Fr k specifies a polynomial P x n = r m codeword length codeword C(x) F n r values of P x evaluated on all possible inputs from F m r.

Reeed Muller code is d + 1-query locally decodable code for any δ, with correctness 1 (d + 1)δ. Decoder For any w Fr m, pick random v Fr m L = {w + λv λ F r }. and consider Query d + 1 codeword positions, corresponding to points of L for λ S, where S F r and S = d + 1. Recover unique univariate polynomial h(λ) ( degree of h is d) If no errors, h(0) = P x (w)

Locally correctable codes C : {0, 1} k {0, 1} n is a (q, δ, ɛ)-lcc if probabilistic A s.t. 1. i [n], x {0, 1} k, y {0, 1} n s.t.d(c(x), y) δn P r[a(y, i) = C(x) i ] 2 1 + ɛ 2. A reads q bits of y Reed Muller codes are locally correctable