MATRIX COMPLETION AND TENSOR RANK

Similar documents
The multiple-vector tensor-vector product

Most tensor problems are NP hard

Sparseness Constraints on Nonnegative Tensor Decomposition

Introduction to Tensors. 8 May 2014

A BLIND SPARSE APPROACH FOR ESTIMATING CONSTRAINT MATRICES IN PARALIND DATA MODELS

Recovering Tensor Data from Incomplete Measurement via Compressive Sampling

Third-Order Tensor Decompositions and Their Application in Quantum Chemistry

CVPR A New Tensor Algebra - Tutorial. July 26, 2017

Fundamentals of Multilinear Subspace Learning

Matrix Completion: Fundamental Limits and Efficient Algorithms

Conditioning and illposedness of tensor approximations

JOS M.F. TEN BERGE SIMPLICITY AND TYPICAL RANK RESULTS FOR THREE-WAY ARRAYS

The convex algebraic geometry of rank minimization

Scalable Tensor Factorizations with Incomplete Data

Exact Low-rank Matrix Recovery via Nonconvex M p -Minimization

Matrix completion: Fundamental limits and efficient algorithms. Sewoong Oh Stanford University

Tutorial on MATLAB for tensors and the Tucker decomposition

New Ranks for Even-Order Tensors and Their Applications in Low-Rank Tensor Optimization

c 2008 Society for Industrial and Applied Mathematics

Eigenvalues of tensors

A concise proof of Kruskal s theorem on tensor decomposition

Approximate Low-Rank Tensor Learning

The Canonical Tensor Decomposition and Its Applications to Social Network Analysis

Window-based Tensor Analysis on High-dimensional and Multi-aspect Streams

Fitting a Tensor Decomposition is a Nonlinear Optimization Problem

Decomposing a three-way dataset of TV-ratings when this is impossible. Alwin Stegeman

Sparse and Low-Rank Matrix Decompositions

Computational Linear Algebra

arxiv: v1 [math.ra] 13 Jan 2009

Tropical decomposition of symmetric tensors

Approximation algorithms for nonnegative polynomial optimization problems over unit spheres

A FLEXIBLE MODELING FRAMEWORK FOR COUPLED MATRIX AND TENSOR FACTORIZATIONS

QUALITATIVE CONTROLLABILITY AND UNCONTROLLABILITY BY A SINGLE ENTRY

A Simpler Approach to Low-Rank Tensor Canonical Polyadic Decomposition

BEYOND MATRIX COMPLETION

TENSOR APPROXIMATION TOOLS FREE OF THE CURSE OF DIMENSIONALITY

Convex envelopes for fixed rank approximation

Recovering any low-rank matrix, provably

TENSOR COMPLETION THROUGH MULTIPLE KRONECKER PRODUCT DECOMPOSITION

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

A Characterization of Sampling Patterns for Union of Low-Rank Subspaces Retrieval Problem

An iterative hard thresholding estimator for low rank matrix recovery

A new method on deterministic construction of the measurement matrix in compressed sensing

Introduction to the Tensor Train Decomposition and Its Applications in Machine Learning

Matrix Completion from a Few Entries

Tensor Rank is Hard to Approximate

A randomized block sampling approach to the canonical polyadic decomposition of large-scale tensors

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

Hyperdeterminants, secant varieties, and tensor approximations

Permutation transformations of tensors with an application

Matrix estimation by Universal Singular Value Thresholding

Tensor Low-Rank Completion and Invariance of the Tucker Core

LIMITATION OF LEARNING RANKINGS FROM PARTIAL INFORMATION. By Srikanth Jagabathula Devavrat Shah

Tensor Decompositions Workshop Discussion Notes American Institute of Mathematics (AIM) Palo Alto, CA

Information-Theoretic Limits of Matrix Completion

THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR

Globally convergent algorithms for PARAFAC with semi-definite programming

Numerical multilinear algebra in data analysis

Multi-Linear Mappings, SVD, HOSVD, and the Numerical Solution of Ill-Conditioned Tensor Least Squares Problems

Uniqueness of the Solutions of Some Completion Problems

MATH 304 Linear Algebra Lecture 10: Linear independence. Wronskian.

To be published in Optics Letters: Blind Multi-spectral Image Decomposition by 3D Nonnegative Tensor Title: Factorization Authors: Ivica Kopriva and A

Homework 2 Foundations of Computational Math 2 Spring 2019

TiDA, winter Pauli Miettinen 1

On the uniqueness of multilinear decomposition of N-way arrays

Problem # Max points possible Actual score Total 120

Fast Angular Synchronization for Phase Retrieval via Incomplete Information

arxiv: v1 [cs.lg] 18 Nov 2018

Robust Principal Component Analysis

Generic properties of Symmetric Tensors

Compressed Sensing and Robust Recovery of Low Rank Matrices

CS168: The Modern Algorithmic Toolbox Lecture #10: Tensors, and Low-Rank Tensor Recovery

Generalized Higher-Order Tensor Decomposition via Parallel ADMM

The University of Texas at Austin Department of Electrical and Computer Engineering. EE381V: Large Scale Learning Spring 2013.

Numerical Multilinear Algebra I

Kronecker Product Approximation with Multiple Factor Matrices via the Tensor Product Algorithm

An Analytic Approach to the Problem of Matroid Representibility: Summer REU 2015

Optimal solutions to non-negative PARAFAC/multilinear NMF always exist

Preliminary/Qualifying Exam in Numerical Analysis (Math 502a) Spring 2012

Linear equations in linear algebra

Optimisation Combinatoire et Convexe.

Properties of Matrices and Operations on Matrices

Tensor Analysis. Topics in Data Mining Fall Bruno Ribeiro

The convex algebraic geometry of linear inverse problems

Tensor Decompositions for Machine Learning. G. Roeder 1. UBC Machine Learning Reading Group, June University of British Columbia

Low-rank Matrix Completion with Noisy Observations: a Quantitative Comparison

Tensors. Notes by Mateusz Michalek and Bernd Sturmfels for the lecture on June 5, 2018, in the IMPRS Ringvorlesung Introduction to Nonlinear Algebra

Quantum Computing Lecture 2. Review of Linear Algebra

An Effective Tensor Completion Method Based on Multi-linear Tensor Ring Decomposition

of Orthogonal Matching Pursuit

Matrix Completion from Fewer Entries

Tensors and graphical models

Linear Algebra March 16, 2019

An Optimization-based Approach to Decentralized Assignability

Row Space and Column Space of a Matrix

arxiv: v1 [math.oc] 11 Jun 2009

Fast Nonnegative Tensor Factorization with an Active-Set-Like Method

Linear dimensionality reduction for data analysis

A completely entangled subspace of maximal dimension

1. Structured representation of high-order tensors revisited. 2. Multi-linear algebra (MLA) with Kronecker-product data.

Transcription:

MATRIX COMPLETION AND TENSOR RANK HARM DERKSEN Abstract. In this paper, we show that the low rank matrix completion problem can be reduced to the problem of finding the rank of a certain tensor. arxiv:1302.2639v2 [math.oc] 23 Jul 2013 1. introduction Suppose we are given a rectangular array which has only been partially filled out with entries in R (or some other field). The low rank matrix completion problem asks to fill out the remaining entries such that the resulting matrix has minimal rank. For example, the array 2 2 3 6 2 can be completed to a rank 1 matrix 2 3 1 2 3 1 4 6 2 but clearly cannot be completed to a rank 0 matrix. For a survey on matrix completion problems, see [13, 11]. The low rank matrix completion problem has various applications, such as statistics, computer vision, signal processing and control. The low rank matrix completion problem is known to be NP-hard (see [16, Theorem 3.1]). For the field F = R or F = C one sometimes can solve the matrix completion problem using convex relaxation: under certain assumptions minimizing the nuclear norm of the matrix(the sum of the singular values) one also obtains the minimal rank completion. See for example [18, 3, 4, 14]. The rank of a tensor was defined in [10]. Using this notion, matrix completion problems can be generalized to higher order tensors (see [6]). Suppose that F is a field, and V (i) = F m i for i = 1,2,...,d. The tensor product space V = V (1) V (2) V (d) can be viewed as a d-dimensional array of size m 1 m 2 m d. A pure tensor is an element of V of the form v (1) v (2) v (d), where v (i) V (i) for all i. The rank of a tensor T V is the smallest nonnegative integer r such that T can be written as a sum of r pure tensors: T = v (1) i v (2) i v (d) i The author was partially supported by NSF grant DMS 0901298. 1

where v (i) j V (i) for all i and j. We denote the tensor rank of T by rank(t). If A = (a i,j ) is an n m matrix, then we can identify A with the tensor a i,j (e i e j ), in F n F m, where e i denotes the i-th basis vector. The tensor rank in this case is exactly the same as the rank of the matrix. For d 3 it is often difficult to find the rank of a given tensor. Finding the rank of a tensor is NP-complete if the field F is finite ([8, 9]) and NP-hard if the field is Q ([8, 9]), or if it contains Q ([12]). Approximation of a tensor by another tensors of low rank is known as the PARAFAC ([7]) or CANDECOMP ([2]) model. There are various algorithms for finding low-rank approximations. See [15, 19] for a discussion. The problems of finding the tensor rank of a tensor, and to approximate tensors with tensors of low rank has many applications, such as the complexity of matrix multiplication, fluorescence spectroscopy, statistics, psychometrics, geophysics and magnetic resonance imaging. In general, the rank of higher tensors more ill-behaved than the rank of a matrix (see [5]). In the next section we will show that the low rank matrix completion problem can be reduced tofinding therankofacertaintensor. Lek-Heng Limpointedout totheauthor, that because low rank matrix completion is NP-hard, this gives another proof that determining the rank of a tensor is NP-hard. 2. The main result Suppose that F is a field, and A = (a i,j ) is an n m array for which some of the entries are given, and the entries in positions (i 1,j 1 ),...,(i s,j s ) are undetermined. Define e i,j as the matrix that has a 1 in position (i,j) and 0 everywhere else. We can reformulate the low rank matrix completion problem as follows. Define a ik,j k = 0 for k = 1,2,...,s and find a matrix B of the form B = A+ λ k e ik,j k, λ 1,...,λ s F. which has minimal rank. We can view matrices as second order tensors. So we identify e i,j with e i e j F n F m and we have A = a i,j (e i e j ). Define u k = e ik and v k = e jk. The matrix completion problem asks to find λ 1,...,λ s F such that the tensor B = A+ λ k (u k v k ) = a i,j (e i e j )+ λ k (u k v k ) has minimal rank. Let U F n F m be the vector space spanned by u 1 v 1,...,u s v s. Definition 1. We define rank(a,u) = min{rank(b) B A+U}. 2

So the matrix completion asks for the value of rank(a,u) and to find a matrix B A+U for which rank(a, U) = rank(b). We define a third order tensor  Fn F m F s+1 by (1)  = A e s+1 + u k v k e k = a i,j (e i e j e s+1 )+ u k v k e k. Our main result is: Theorem 2. We have rank(a,u)+s = rank(â). The theorem is effective: given an explicit decomposition of  as a sum of l = rank(â) pure tensors, the proof shows that one easily can construct a matrix B in A+U such that rank(b) = l s = rank(a,u). Proof of Theorem 2. Let r = rank(a,u) and l = rank(â). By definition, there exists a matrix B A+U with rank r. We can write B = A+ λ k (u k v k ). Since B has rank r, we also can write B = x i y i for some x 1,...,x r F n and y 1,...,y r F m. We have  = A e s+1 + = B e s+1 + u k v k e k = (B λ k u k v k ) e s+1 + u k v k (e k λ k e s+1 ) = x i y i + We have written  as a sum of r+s pure tensors. So l = rank(â) r +s. Conversely, we can write  = a j b j c j u k v k e k = u k v k (e k λ k e s+1 ). For every i, choose a linear function h i : F n F m F such that h i (u i v i ) = 1 and h i (u j v j ) = 0 if j i. Applying h i id : F n F m F s+1 F s+1 to the tensor  gives h i (a j b j )c j = (h i id)(â) = e i +h i (A)e s+1. This shows that e i lies in the span of c 1,...,c l,e s+1 for all i. So c 1,...,c l,e s+1 span the vector space F s+1. After rearranging c 1,...,c l, we may assume that c 1,c 2,...,c s,e s+1 is a 3

basis of F s+1. There exists a linear function f : F s+1 F such that f(c 1 ) = = f(c s ) = 0 and f(e s+1 ) = 1. We apply id id f : F n F m F s+1 F n F m to Â: A+ f(e j )u i v i = (id id f)(â) = f(c j )a j b j = f(c j )a j b j. j=s+1 The tensor above is a matrix in A+U which has rank at most l s, and by definition of r = rank(a,u) it has rank at least r. It follows that r l s. We have already proven that r +s l, so we conclude that r +s = l. Remark 3. In the proof we have not used that u k and v k are basis vectors. So Theorem 2 is true whenever U is the span of linearly independent pure tensors and  is given by (1). u 1 v 1,...,u s v s Remark 4. Define t i (F 2 ) s = } F 2 F {{} 2 s by t i = e 1 e 1 e 2 e 1 e 1 where e 2 appears in the i-th position if i s, and We define a tensor à Fn F m (F 2 ) s by t s+1 = e 1 e 1 e 1. à = u 1 v 1 t 1 + +u s v s t s +A t s+1. A proof, similar to that of Theorem 2 shows that rank(ã) = s+rank(a,u). The ambient vector space of à is much larger than that of Â. But one can imagine that using the tensor à can be advantageous for proving lower bounds for rank(a,u) because it is, in a way, more rigid. Remark 5. Theorem 2 also easily generalizes to the higher order analog of matrix completion: tensor completion. Suppose that V (1),...,V (d) are finite dimensional F-vector spaces and V = V (1) V (d). Assume that U V is a subspace spanned by s linearly independent pure tensors and is a tensor. As before, we define u (1) i u (2) i u (d) i, i = 1,2,...,s A V (1) V (d) rank(a,u) = min{rank(b) B A+U}. Define a (d+1)-th order tensor by  = A e s+1 + u (1) k u (d) k e k V (1) V (d) F s+1. 4

Then we have rank(a,u)+s = rank(â). Acknowledgment. The author thanks Lek-Heng Lim for some useful suggestions, and Reghu Meka for providing me with an important reference. References [1] E. Acar, B. Yener, Unsupervised multiway data analysis: a literature study, IEEE Trans. on Knowledge and Data Engin. 21 (2009), 6 20. [2] J. D. Carroll and J. Chang, Analysis of individual differences in multidimensional scaling via an n-way generalization of Eckart-Young decomposition, Psychometrika 35 (1970), 283 319. [3] E.J. Candès and B. Recht, Exact matrix completion via convex optimization, Foundations of Computational Mathematics 9, 2009, 717 772. [4] E. J. Candès and T. Tao, The power of convex relaxation: Near-optimal matrix completion, Preprint. [5] V. De Silva and L.-H. Lim, Tensor rank and the ill-posedness of the best low-rank approximation problem, SIAM J. Matrix Analysis Appl. 30 (2008), no. 3, 1084 1127. [6] S. Gandy, B. Recht and I. Yamada, Tensor completion and low n-rank tensor recovery via convex optimization, Inverse Problems 27 (2011), no. 2. [7] R. A. Harshman, Foundations of the parafac procedure: models and conditions for an explanatory multimodal factor analysis, UCLA working papers in Phonetics 16 (1970), 1 84. [8] J. Håstad, Tensor rank is NP-complete, J. Algorithms 11 (1990), no. 4, 644 654. [9] J. Håstad, Tensor rank is NP-complete, Automata, languages and programming (Stresa, 1989), Lecture Notes in Comput. Sci. 372, Springer, Berlin, 1989, 451 460. [10] F. L. Hitchcock, Multiple invariants and generalized rank of a p- way matrix or tensor, J. Math. and Phys. 7 (1927), no. 1, 40 79. [11] M. Laurent, Matrix completion problems, in: The Encyclopedia of Optimization III, C. A. Floudas and P..M. Pardalos, eds., Kluwer, 2001. [12] C. Hillar and L.-H. Lim, Most tensor problems are NP-hard, Journal of the ACM (2013), to appear. [13] C. R. Johnson, Matrix completion problems: a survey, in: Matrix Theory and Applications, AMS, Providence, RI, 1990, 171 198. [14] R. H. Keshavan, A. Montanari and S. Oh, Matrix completion from a few entries, IEEE Trans. on Information Theory 56 (2010), 2980 2998. [15] T. G. Kolda, B. W. Bader, Tensor Decompositions and Applications, SIAM Rev. 51 (3), 455 500. [16] R. Peeters, Orthogonal representations over finite fields and the chromatic number of graphs, Combinatorica 16 (1996), 417 431. [17] B. Recht, A simpler approach to matrix competion, J. Mach. Learn. Res. 12 (2011), 3413 3430. [18] B. Recht, M. Fazel and P. Parillo, Guaranteed minimum rank solutions of matrix equations via nuclear norm minimization, SIAM Review 52 (2010), no. 3, 471 501. [19] G. Tomasi and R. Bro, A comparison of algorithms for fitting the parafac model, Computational Statistics and Data Analysis 50, 7, 1700 1734. Harm Derksen Department of Mathematics University of Michigan 530 Church Street Ann Arbor, MI 48109-1043, USA hderksen@umich.edu 5