An Approximation Algorithm for Approximation Rank

Similar documents
Rank minimization via the γ 2 norm

Approximation norms and duality for communication complexity lower bounds

An approximation algorithm for approximation rank

An approximation algorithm for approximation rank

Direct product theorem for discrepancy

Direct product theorem for discrepancy

Lower Bounds in Communication Complexity

Grothendieck Inequalities, XOR games, and Communication Complexity

Direct product theorem for discrepancy

A direct product theorem for discrepancy

15-251: Great Theoretical Ideas In Computer Science Recitation 9 : Randomized Algorithms and Communication Complexity Solutions

Matrix Rank in Communication Complexity

Spectral gaps and geometric representations. Amir Yehudayoff (Technion)

Lower Bounds in Communication Complexity Based on Factorization Norms

Approximating sparse binary matrices in the cut-norm

Rank, Trace-Norm & Max-Norm

Lecture 7: Positive Semidefinite Matrices

be a Householder matrix. Then prove the followings H = I 2 uut Hu = (I 2 uu u T u )u = u 2 uut u

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2

Section 3.9. Matrix Norm

Breaking the Rectangle Bound Barrier against Formula Size Lower Bounds

Communication is bounded by root of rank

Communication Complexity

Partitions and Covers

The quantum adversary method and classical formula size lower bounds

Optimal compression of approximate Euclidean distances

CS286.2 Lecture 8: A variant of QPCP for multiplayer entangled games

Simultaneous Diagonalization of Positive Semi-definite Matrices

A Unified Theorem on SDP Rank Reduction. yyye

On Quantum vs. Classical Communication Complexity

The Graph Realization Problem

Nondeterminism LECTURE Nondeterminism as a proof system. University of California, Los Angeles CS 289A Communication Complexity

Multi-stage convex relaxation approach for low-rank structured PSD matrix recovery

Some upper and lower bounds on PSD-rank

Product theorems via semidefinite programming

An exponential separation between quantum and classical one-way communication complexity

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2017 LECTURE 5

Ma/CS 6b Class 20: Spectral Graph Theory

The Alon-Saks-Seymour and Rank-Coloring Conjectures

Chapter 1. Matrix Algebra

ELE/MCE 503 Linear Algebra Facts Fall 2018

Adaptive Matrix Vector Product

Lecture 8: Linear Algebra Background

Bounded Matrix Rigidity and John s Theorem

The Computational Complexity Column

Random Methods for Linear Algebra

Parallel Singular Value Decomposition. Jiaxing Tan

Knowledge Discovery and Data Mining 1 (VO) ( )

STRONG DIRECT PRODUCT THEOREMS FOR QUANTUM COMMUNICATION AND QUERY COMPLEXITY

Linear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,

Approximating Submodular Functions. Nick Harvey University of British Columbia

On the exponential convergence of. the Kaczmarz algorithm

Lecture 15 Review of Matrix Theory III. Dr. Radhakant Padhi Asst. Professor Dept. of Aerospace Engineering Indian Institute of Science - Bangalore

Distributed Statistical Estimation of Matrix Products with Applications

Rank Determination for Low-Rank Data Completion

On Delay-Independent Diagonal Stability of Max-Min Min Congestion Control

Lecture 2: Linear operators

STAT 309: MATHEMATICAL COMPUTATIONS I FALL 2013 PROBLEM SET 2

9. Numerical linear algebra background

Tensor product Take two tensors, get together enough inputs to feed into both, and take the product of their results.

Lecture 18. Ramanujan Graphs continued

Ma/CS 6b Class 20: Spectral Graph Theory

An Introduction to Linear Matrix Inequalities. Raktim Bhattacharya Aerospace Engineering, Texas A&M University

Lecture 1: Review of linear algebra

LinGloss. A glossary of linear algebra

Section 3.3. Matrix Rank and the Inverse of a Full Rank Matrix

arxiv: v1 [cs.cc] 12 Feb 2009

Optimal quantum adversary lower bounds for ordered search

arxiv: v3 [math.ra] 10 Jun 2016

OPERATIONS on large matrices are a cornerstone of

Z-Pencils. November 20, Abstract

Introduction to Compressed Sensing

Functional Analysis Review

arxiv: v1 [math.fa] 19 Jul 2009

On Convex Complexity Measures

Efficient quantum protocols for XOR functions

EE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 4

Some Selected Topics. Spectral Norm in Learning Theory:

Lecture 3. Random Fourier measurements

A Tutorial on Matrix Approximation by Row Sampling

A lower bound for the Laplacian eigenvalues of a graph proof of a conjecture by Guo

Conditions for Robust Principal Component Analysis

Data Mining and Matrices

Principal Component Analysis

New Characterizations in Turnstile Streams with Applications

Lecture # 3 Orthogonal Matrices and Matrix Norms. We repeat the definition an orthogonal set and orthornormal set.

Quantum entanglement, sum of squares, and the log rank conjecture.

Dichotomy Theorems for Counting Problems. Jin-Yi Cai University of Wisconsin, Madison. Xi Chen Columbia University. Pinyan Lu Microsoft Research Asia

Lecture 16: Communication Complexity

Randomized Communication Complexity for Linear Algebra Problems over Finite Fields

Introduction to Numerical Linear Algebra II

25.2 Last Time: Matrix Multiplication in Streaming Model

Convex sets, conic matrix factorizations and conic rank lower bounds

CHAPTER 11. A Revision. 1. The Computers and Numbers therein

Network Localization via Schatten Quasi-Norm Minimization

The University of Texas at Austin Department of Electrical and Computer Engineering. EE381V: Large Scale Learning Spring 2013.

Ir O D = D = ( ) Section 2.6 Example 1. (Bottom of page 119) dim(v ) = dim(l(v, W )) = dim(v ) dim(f ) = dim(v )

Lecture: Expanders, in theory and in practice (2 of 2)

Throughout these notes we assume V, W are finite dimensional inner product spaces over C.

Transcription:

An Approximation Algorithm for Approximation Rank Troy Lee Columbia University Adi Shraibman Weizmann Institute

Conventions Identify a communication function f : X Y { 1, +1} with the associated X-by-Y matrix A(x,y)=f(x,y). Denote by D(A) the deterministic communication complexity of the sign matrix A. Denote by R ɛ (A) the randomized complexity with error at most ɛ.

How a protocol partitions a matrix Bob Y 0 Alice X 1

How a protocol partitions a matrix Bob Y 00 01 Alice X 11 10

How a protocol partitions a matrix Bob Y Alice X 001 000 11 01 100 101

How a protocol partitions a matrix Bob Y Alice X 000 001 011 010 110 111 110 111

Log rank bound

A c Log rank bound bit protocol partitions the communication matrix into at most 2 c many rectangles.

A c Log rank bound bit protocol partitions the communication matrix into at most 2 c many rectangles. If the protocol is correct, each rectangle will be monochromatic, and so rank one.

A c Log rank bound bit protocol partitions the communication matrix into at most 2 c many rectangles. If the protocol is correct, each rectangle will be monochromatic, and so rank one. By subadditivity of rank, this gives D(A) log rank(a)

A c Log rank bound bit protocol partitions the communication matrix into at most 2 c many rectangles. If the protocol is correct, each rectangle will be monochromatic, and so rank one. By subadditivity of rank, this gives D(A) log rank(a) One of the greatest open problems in communication complexity, the log rank conjecture of Lovasz and Saks, states that this bound is polynomially tight

Approximation rank

Approximation rank For randomized (and quantum) models, one can show that a lower bound is given by approximation rank. [Krause96, Buhrman and de Wolf 01]

Approximation rank For randomized (and quantum) models, one can show that a lower bound is given by approximation rank. [Krause96, Buhrman and de Wolf 01] Given a target matrix A, approximation rank looks at the minimal rank matrix entrywise close to A: rank ɛ (A) = min X rank(x) X(i, j) A(i, j) ɛ for all i, j.

Approximation rank

Approximation rank In analogy with the log rank conjecture, can also conjecture that log approximation rank is polynomially tight for randomized communication complexity.

Approximation rank In analogy with the log rank conjecture, can also conjecture that log approximation rank is polynomially tight for randomized communication complexity. We do not know if approximation rank is NP-hard to compute, but similar rank minimization problems are NP-hard.

Approximation rank In analogy with the log rank conjecture, can also conjecture that log approximation rank is polynomially tight for randomized communication complexity. We do not know if approximation rank is NP-hard to compute, but similar rank minimization problems are NP-hard. While approximation rank gives a lower bound on quantum communication complexity, not known to work for model with entanglement.

Main result

Main result We show a semidefinite programming quantity which is polynomially related to approximation rank for constant 0 < ε< 1/2

Main result We show a semidefinite programming quantity which is polynomially related to approximation rank for constant 0 < ε< 1/2 Moreover, this quantity is known to be a lower bound even in the quantum model with entanglement.

Convex relaxation

Convex relaxation Part of what makes rank minimization problems hard is that rank is not a convex function.

Convex relaxation Part of what makes rank minimization problems hard is that rank is not a convex function. As is common in approximation algorithms, we look at a convex relaxation of the problem.

Convex relaxation Part of what makes rank minimization problems hard is that rank is not a convex function. As is common in approximation algorithms, we look at a convex relaxation of the problem. For this, we need to introduce some matrix norms.

Matrix norms Define the i th singular value as Denote σ i (A) = λ i (AA t ) A 1 = i σ i (A) A = σ 1 (A) A 2 = σi 2(A) i

Matrix norms Define the i th singular value as Denote σ i (A) = λ i (AA t ) A 1 = i σ i (A) Trace Norm A = σ 1 (A) A 2 = σi 2(A) i

Matrix norms Define the i th singular value as Denote σ i (A) = λ i (AA t ) A 1 = i σ i (A) Trace Norm A = σ 1 (A) A 2 = σi 2(A) i Spectral Norm

Matrix norms Define the i th singular value as Denote σ i (A) = λ i (AA t ) A 1 = i σ i (A) Trace Norm A = σ 1 (A) A 2 = σi 2(A) i Spectral Norm Frobenius Norm

Matrix norms Define the i th singular value as Denote σ i (A) = λ i (AA t ) A 1 = i σ i (A) Trace Norm A = σ 1 (A) A 2 = σi 2(A) i Spectral Norm Frobenius Norm = A(i, j) 2 i,j

Trace norm method Replace the rank objective function by the trace norm. As rank equals the number of nonzero singular values, we have i σ i (A) rank(a) A 2 For a M-by-N sign matrix this gives rank(a) A 2 1 MN

Example: Trace Norm Let H be a N-by-N Hadamard matrix.

Example: Trace Norm Let H be a N-by-N Hadamard matrix. Then HH t = NI and so H 1 = N 3/2

Example: Trace Norm Let H be a N-by-N Hadamard matrix. Then HH t = NI and so H 1 = N 3/2 Thus trace norm method gives a bound on the rank of H 2 1 H 2 2 = N 3 N 2 = N

Trace norm method: drawback Trace norm bound is not monotone. Consider ( HN 1 N 1 N 1 N )

Trace norm method: drawback Trace norm bound is not monotone. Consider ( HN 1 N 1 N 1 N ) Trace norm is at most N 3/2 +3N

Trace norm method: drawback Trace norm bound is not monotone. Consider ( HN 1 N 1 N 1 N ) Trace norm is at most N 3/2 +3N Trace norm method gives (N 3/2 +3N) 2 4N 2 = N 4 + O( N)

Trace norm method: drawback Trace norm bound is not monotone. Consider ( HN 1 N 1 N 1 N ) Trace norm is at most N 3/2 +3N Trace norm method gives (N 3/2 +3N) 2 4N 2 = N 4 + O( N) Worse bound than on Hadamard submatrix!

A fix We can fix this by considering max u,v u = v =1 A uv t 1 As this entrywise product does not increase the rank we still have ( A uv t 1 ) 2 rank(a) A uv t 2 For a sign matrix A, this simplifies nicely: rank(a) ( A uv t 1 ) 2

We have arrived At the γ 2 norm, introduced in communication complexity by [LMSS07, LS07]. γ 2 (A) = max u,v u = v =1 A uv t 1

We have arrived At the γ 2 norm, introduced in communication complexity by [LMSS07, LS07]. γ 2 (A) = max u,v u = v =1 A uv t 1 Known by a variety of names in matrix analysis: Hadamard/Schur product operator/trace norm

We have arrived At the γ 2 norm, introduced in communication complexity by [LMSS07, LS07]. γ 2 (A) = max u,v u = v =1 A uv t 1 Known by a variety of names in matrix analysis: Hadamard/Schur product operator/trace norm Has many nice properties. For this talk, we use the fact that it can be written as a semidefinite program and computed to high accuracy in polynomial time.

Application to approximation rank

Application to approximation rank Use γ 2 norm as surrogate in rank minimization problem γ ɛ 2(A) = min B γ 2(B) A(i, j) B(i, j) ɛ for all i, j.

Application to approximation rank Use γ 2 norm as surrogate in rank minimization problem γ ɛ 2(A) = min B γ 2(B) A(i, j) B(i, j) ɛ for all i, j. As argued above, γ ɛ 2(A) 2 (1 + ɛ) 2 rank ɛ(a).

γ 2 norm as factorization l 2 norm l 2 norm A = X Y γ 2 (A) = min X,Y XY =A X r Y c

Rank as factorization k k Y A = X

γ 2 norm as factorization l 2 norm l 2 norm A = X Y γ 2 (A) = min X,Y XY =A X r Y c

Now shrink the rows Take X,Y optimal factorization realizing γ ɛ 2(A). Obtain matrices X,Y by randomly projecting columnspace down to dimension about γ ɛ 2(A) 2 log N γ ɛ 2(A) 2 log N X = X R

Johnson-Lindenstrauss Lemma Now X Y will be a matrix of the desired rank. Furthermore, by the Johnson-Lindenstrauss lemma and a union bound, the inner products between all rows of X and columns of Y will approximately equal those between X and Y. Thus X Y will still be entrywise close to A.

Error-reduction Can argue that if we started with an ε approximation, now will have 2ε approximation. Can fix this by applying low degree polynomial approximation of the sign function entrywise to the matrix. Applying a degree d polynomial blows up rank by at most a power of d.

Error-reduction (7x-x^3)/6 1.6 1.2 0.8 0.4-2.4-2 -1.6-1.2-0.8-0.4 0 0.4 0.8 1.2 1.6 2 2.4-0.4-0.8-1.2

Final result For any M-by-N sign matrix A and constant 0 < ε< 1 γ ɛ 2(A) 2 (1 + ɛ) 2 rank ɛ(a) =O ( γ ɛ 2(A) 2 log(mn) ) 3 Implies that the log approximation rank conjecture is essentially equivalent to the existence of a constant c such that R ɛ (A) (log γ ɛ 2(A)) c + O(log n).

Open questions What is the complexity of the real vector inner product function? Does the log rank conjecture imply the log approximation rank conjecture? Approximation algorithm for the limiting case of sign rank?