CVPR A New Tensor Algebra - Tutorial. July 26, 2017

Similar documents
Novel methods for multilinear data completion and de-noising based on tensor-svd

Tensor-Tensor Product Toolbox

ENGG5781 Matrix Analysis and Computations Lecture 10: Non-Negative Matrix Factorization and Tensor Decomposition

Computational Linear Algebra

FACTORIZATION STRATEGIES FOR THIRD-ORDER TENSORS

Low-Rank Tensor Completion by Truncated Nuclear Norm Regularization

Parallel Tensor Compression for Large-Scale Scientific Data

Fundamentals of Multilinear Subspace Learning

Faloutsos, Tong ICDE, 2009

From Matrix to Tensor. Charles F. Van Loan

N-mode Analysis (Tensor Framework) Behrouz Saghafi

Main matrix factorizations

Introduction to Tensors. 8 May 2014

MATRIX COMPLETION AND TENSOR RANK

Third-Order Tensor Decompositions and Their Application in Quantum Chemistry

Tensor Decompositions for Machine Learning. G. Roeder 1. UBC Machine Learning Reading Group, June University of British Columbia

A new truncation strategy for the higher-order singular value decomposition

c Springer, Reprinted with permission.

The multiple-vector tensor-vector product

Multilinear Subspace Analysis of Image Ensembles

Low-tubal-rank Tensor Completion using Alternating Minimization

Tensor Decompositions and Applications

A Tensor Approximation Approach to Dimensionality Reduction

Probabilistic Latent Semantic Analysis

SPARSE TENSORS DECOMPOSITION SOFTWARE

c 2008 Society for Industrial and Applied Mathematics

Tensor Robust Principal Component Analysis: Exact Recovery of Corrupted Low-Rank Tensors via Convex Optimization

To be published in Optics Letters: Blind Multi-spectral Image Decomposition by 3D Nonnegative Tensor Title: Factorization Authors: Ivica Kopriva and A

/16/$ IEEE 1728

Lecture 4. Tensor-Related Singular Value Decompositions. Charles F. Van Loan

Slice Oriented Tensor Decomposition of EEG Data for Feature Extraction in Space, Frequency and Time Domains

Higher-Order Singular Value Decomposition (HOSVD) for structured tensors

Multilinear Analysis of Image Ensembles: TensorFaces

THIRD ORDER TENSORS AS OPERATORS ON MATRICES: A THEORETICAL AND COMPUTATIONAL FRAMEWORK WITH APPLICATIONS IN IMAGING

Tensor Analysis. Topics in Data Mining Fall Bruno Ribeiro

Sparseness Constraints on Nonnegative Tensor Decomposition

Decomposing a three-way dataset of TV-ratings when this is impossible. Alwin Stegeman

arxiv: v1 [cs.cv] 7 Jan 2019

Principal Component Analysis

TENSOR LAYERS FOR COMPRESSION OF DEEP LEARNING NETWORKS. Cris Cecka Senior Research Scientist, NVIDIA GTC 2018

Data Mining and Matrices

Postgraduate Course Signal Processing for Big Data (MSc)

Linear Subspace Models

Multiscale Tensor Decomposition

PRINCIPAL Component Analysis (PCA) is a fundamental approach

Rank Determination for Low-Rank Data Completion

A Multi-Affine Model for Tensor Decomposition

PAVEMENT CRACK CLASSIFICATION BASED ON TENSOR FACTORIZATION. Offei Amanor Adarkwa

Must-read Material : Multimedia Databases and Data Mining. Indexing - Detailed outline. Outline. Faloutsos

Tensor Decompositions Workshop Discussion Notes American Institute of Mathematics (AIM) Palo Alto, CA

2.3. Clustering or vector quantization 57

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

Kronecker Product Approximation with Multiple Factor Matrices via the Tensor Product Algorithm

B553 Lecture 5: Matrix Algebra Review

Rank Selection in Low-rank Matrix Approximations: A Study of Cross-Validation for NMFs

Principal Component Analysis (PCA)

Introduction to Machine Learning

Fitting a Tensor Decomposition is a Nonlinear Optimization Problem

CPSC 340: Machine Learning and Data Mining. More PCA Fall 2017

Scalable Tensor Factorizations with Incomplete Data

Theoretical Performance Analysis of Tucker Higher Order SVD in Extracting Structure from Multiple Signal-plus-Noise Matrices

JOS M.F. TEN BERGE SIMPLICITY AND TYPICAL RANK RESULTS FOR THREE-WAY ARRAYS

Tutorial on MATLAB for tensors and the Tucker decomposition

Nonnegative Tensor Factorization with Smoothness Constraints

CP DECOMPOSITION AND ITS APPLICATION IN NOISE REDUCTION AND MULTIPLE SOURCES IDENTIFICATION

Lecture: Face Recognition and Feature Reduction

A Simpler Approach to Low-Rank Tensor Canonical Polyadic Decomposition

An Effective Tensor Completion Method Based on Multi-linear Tensor Ring Decomposition

Kronecker Decomposition for Image Classification

1 Singular Value Decomposition and Principal Component

arxiv: v1 [math.na] 22 Sep 2016

Three right directions and three wrong directions for tensor research

Automatic Rank Determination in Projective Nonnegative Matrix Factorization

Online Tensor Factorization for. Feature Selection in EEG

Window-based Tensor Analysis on High-dimensional and Multi-aspect Streams

Linear Algebra (Review) Volker Tresp 2017

TENSOR COMPLETION VIA ADAPTIVE SAMPLING OF TENSOR FIBERS: APPLICATION TO EFFICIENT INDOOR RF FINGERPRINTING

Performance Evaluation of the Matlab PCT for Parallel Implementations of Nonnegative Tensor Factorization

Example: Face Detection

An Online Tensor Robust PCA Algorithm for Sequential 2D Data

Lecture: Face Recognition and Feature Reduction

A randomized block sampling approach to the canonical polyadic decomposition of large-scale tensors

Linear Algebra (Review) Volker Tresp 2018

Coupled Matrix/Tensor Decompositions:

The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA)

Robust Tensor Completion and its Applications. Michael K. Ng Department of Mathematics Hong Kong Baptist University

High Order Tensor Formulation for Convolutional Sparse Coding

Face Recognition. Face Recognition. Subspace-Based Face Recognition Algorithms. Application of Face Recognition

Lecture 1: Introduction to low-rank tensor representation/approximation. Center for Uncertainty Quantification. Alexander Litvinenko

Matrix-Tensor and Deep Learning in High Dimensional Data Analysis

A Tensor Completion Approach for Efficient and Robust Fingerprint-based Indoor Localization

Non-negative Matrix Factor Deconvolution; Extraction of Multiple Sound Sources from Monophonic Inputs

A Modular NMF Matching Algorithm for Radiation Spectra

Orthogonal tensor decomposition

PCA FACE RECOGNITION

Data Mining and Matrices

Matrices and Vectors

THERE is an increasing need to handle large multidimensional

Lecture Notes 10: Matrix Factorization

The Canonical Tensor Decomposition and Its Applications to Social Network Analysis

Transcription:

CVPR 2017 A New Tensor Algebra - Tutorial Lior Horesh lhoresh@us.ibm.com Misha Kilmer misha.kilmer@tufts.edu July 26, 2017

Outline Motivation Background and notation New t-product and associated algebraic framework Implementation considerations The t-svd and optimality Application in Facial Recognition Proper Orthogonal Decomposition - Dynamic Model Redcution A tensor Nuclear Norm from the t-svd Applications in video completion CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 2

Tensor Applications: Machine vision: understanding the world in 3D, enable understanding phenomena such as perspective, occlusions, illumination Latent semantic tensor indexing: common terms vs. entries vs. parts, co-occurrence of terms Tensor subspace Analysis for Viewpoint Recognition, T. Ivanov, L. Mathies, M.A.O. Vasilescu, ICCV, 2nd IEEE International Workshop on Subspace Methods, September, 2009 CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 3

Tensor Applications: Medical imaging: naturally involves 3D (spatio) and 4D (spatio-temporal) correlations Video surveillance and Motion signature: 2D images + 3 rd dimension of time, 3D/4D motion trajectory Multi-target Tracking with Motion Context in Tenor Power Iteration X. Shi, H. Ling, W. Hu, C. Yuan, and J. Xing IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), Columbus OH, 2014 CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 4

Tensors: Historical Review 1927 F.L. Hitchcock: The expression of a tensor or a polyadic as a sum of products (Journal of Mathematics and Physics) 1944 R.B. Cattell introduced a multiway model: Parallel proportional profiles and other principles for determining the choice of factors by rotation (Psychometrika) F.L. Hitchcock 1960 L.R. Tucker: Some mathematical notes on three-mode factor analysis (Psychometrika) 1981 tensor decomposition was first used in chemometrics Past decade, computer vision, image processing, data mining, graph analysis, etc. R.B. Cattell L.R. Tucker CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 5

The Power of Proper Representation What is that? CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 6

The Power of Proper Representation What is that? Let s observe the same data but in a different (matrix rather than vector) representation CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 7

The Power of Proper Representation Representation matters! some correlations can only be realized in appropriate representation CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 8 What is that? Let s observe the same data but in a different (matrix rather than vector) representation

Motivation Much real-world data is inherently multidimensional color video data 4 way 3D medical image, evolving in time (4 way); multiple patients (5 way) Many operators and models are also multi-way Traditional matrix-based methods based on data vectorization (e.g. matrix PCA) generally agnostic to possible high dimensional correlations Can we uncover hidden patterns in tensor data by computing an appropriate tensor decomposition/approximation? Need to decide on the tensor decomposition application dependent! What do we mean by decompose? CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 9

Tensors: Background and Notation Notation : A n1 n2..., nj - j th order tensor Examples CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 10

Tensors: Background and Notation Notation : A n1 n2..., nj - j th order tensor Examples 0 th order tensor CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 11

Tensors: Background and Notation Notation : A n1 n2..., nj - j th order tensor Examples 0 th order tensor - scalar CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 12

Tensors: Background and Notation Notation : A n1 n2..., nj - j th order tensor Examples 0 th order tensor - scalar 1 st order tensor CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 13

Tensors: Background and Notation Notation : A n1 n2..., nj - j th order tensor Examples 0 th order tensor - scalar 1 st order tensor - vector CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 14

Tensors: Background and Notation Notation : A n1 n2..., nj - j th order tensor Examples 0 th order tensor - scalar 1 st order tensor - vector 2 nd order tensor CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 15

Tensors: Background and Notation Notation : A n1 n2..., nj - j th order tensor Examples 0 th order tensor - scalar 1 st order tensor - vector 2 nd order tensor - matrix CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 16

Tensors: Background and Notation Notation : A n1 n2..., nj - j th order tensor Examples 0 th order tensor - scalar 1 st order tensor - vector 2 nd order tensor - matrix 3 rd order tensor... CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 17

Tensors: Background and Notation Notation : A n1 n2..., nj - j th order tensor Examples 0 th order tensor - scalar 1 st order tensor - vector 2 nd order tensor - matrix 3 rd order tensor... CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 18

Notation A i,j,k = element of A in row i, column j, tube k CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 19

Notation A i,j,k = element of A in row i, column j, tube k A 4,7,1 CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 20

Notation A i,j,k = element of A in row i, column j, tube k A 4,7,1 A :,3,1 CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 21

Notation A i,j,k = element of A in row i, column j, tube k A 4,7,1 A :,3,1 A :,:,3 CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 22

Tensors: Background and Notation Fiber - a vector defined by fixing all but one index while varying the rest Slice - a matrix defined by fixing all but two indices while varying the rest CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 23

Tensor Multiplication Definition : The k - mode multiplication of a tensor X R n1 n2..., n d with a matrix U R J n k is denoted by X k U and is of size n 1 n k 1 J n k+1 n d Element-wise (X k U) i1 i k 1 ji k+1 i d = n d i k =1 x i1i 2 i d u jik 1-mode multiplication CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 24

Tensor Holy Grail and the Matrix Analogy Find a way to express a tensor that leads to the possibility for compressed representation (near redundancy removed) that maintains important features of the original tensor min A B F s.t. B has rank p r B = p i=1 σ i( V u (i) V v (i) ) where A = r i=1 σ i(u (i) v (i) ) CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 25

Tensor Decompositions - CP CP (CANDECOMP-PARAFAC) Decomposition 1 : Outer product r X a i b i c i i=1 T = u v w T ijk = u i v j w k Columns of A = [a 1,..., a r], B = [b 1,..., b r], C = [c 1,..., c r] are not orthogonal If r is minimal, then r is called the rank of the tensor No perfect procedure for fitting CP for a given number of components 2 1 R. Harshman, 1970; J. Carroll and J. Chang, 1970 2 V. de Silva, L. Lim, Tensor Rank and the Ill-Posedness of the Best Low-Rank Approximation Problem, 2008 CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 26

Tensor Decompositions - Tucker Tucker Decomposition : X C 1 G 2 T 3 S = r 1 r 2 r 3 i=1 j=1 k=1 c ijk g i t j s k C is the core tensor G, T, S are the components of factors Can either have diagonal core or orthogonal columns in components [DeLathauwer et al.] Truncated Tucker decomposition is not optimal in approximating the norm of the difference X C 1 G 2 T 3 S CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 27

Tensor Decompositions - t-product t-product : Let A be n 1 n 2 n 3 and B be n 2 l n 3. Then the t-product A B is the n 1 l n 3 tensor A B = fold(circ(a) vec(b)) circ (A) vec (B) = A 1 A n3 A n3 1 A 2 A 2 A 1 A n3 A 3........... A n3 1 A n3 2 A n3 3 A n3 A n3 A n3 1 A n3 2 A 1 B 1 B 2 B 3. B n3 fold(vec(b)) = B A i, B i, i = 1,..., n 3 are frontal slices of A and B M.E. Kilmer and C.D. Martin. Factorization strategies for third-order tensors, Linear Algebra and its Applications, Special Issue in Honor of G. W. Stewart s 70 th birthday, vol. 435(3):641 658, 2011 CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 28

Block Circulants A block circulant can be block-diagonalized by a (normalized) DFT in the 2 nd dimension: Â 1 0 0 (F I)circ (A) (F 0 Â 2 0 I) =. 0.. 0 0 0 Â n Here is a Kronecker product of matrices If F is n n, and I is m m, (F I) is the mn mn block matrix, of n block rows and columns, each block is m m, where the ij th block is f i,j I But we never implement it this way because an FFT along tube fibers of A yields a tensor, Â whose frontal slices are the Âi CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 29

t-product Identity Definition: The n n l identity tensor I nnl is the tensor whose frontal face is the n n identity matrix, and whose other faces are all zeros CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 30

t-product Identity Definition: The n n l identity tensor I nnl is the tensor whose frontal face is the n n identity matrix, and whose other faces are all zeros Class Exercise: Let A be n 1 n n 3, show that A I = A and I A = A CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 31

t-product Identity Definition: The n n l identity tensor I nnl is the tensor whose frontal face is the n n identity matrix, and whose other faces are all zeros Class Exercise: Let A be n 1 n n 3, show that A I = A and I A = A A I CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 32

t-product Identity Definition: The n n l identity tensor I nnl is the tensor whose frontal face is the n n identity matrix, and whose other faces are all zeros Class Exercise: Let A be n 1 n n 3, show that A I = A and I A = A A I = fold (circ (A) vec (I)) CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 33

t-product Identity Definition: The n n l identity tensor I nnl is the tensor whose frontal face is the n n identity matrix, and whose other faces are all zeros Class Exercise: Let A be n 1 n n 3, show that A I = A and I A = A A I = fold (circ (A) vec (I)) = A 1 A n3 A n3 1 A 2 A 2 A 1 A n3 A 3........... A n3 1 A n3 2 A n3 3 A n3 A n3 A n3 1 A n3 2 A 1 I 0 0. 0 CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 34

t-product Identity Definition: The n n l identity tensor I nnl is the tensor whose frontal face is the n n identity matrix, and whose other faces are all zeros Class Exercise: Let A be n 1 n n 3, show that A I = A and I A = A A I = fold (circ (A) vec (I)) = A 1 A n3 A n3 1 A 2 A 2 A 1 A n3 A 3........... A n3 1 A n3 2 A n3 3 A n3 A n3 A n3 1 A n3 2 A 1 I 0 0. 0 = A CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 35

t-product Transpose Definition: If A is n 1 n 2 n 3, then A is the n 2 n 1 n 3 tensor obtained by transposing each of the frontal faces and then reversing the order of transposed faces 2 through n 3 Example: If A R n1 n2 4 and its frontal faces are given by the n 1 n 2 matrices A 1, A 2, A 3, A 4, then A 1 A = fold A 4 A 3 A 2 Mimetic property: when n = 1, the operator collapses to traditional matrix multiplication between two matrices and tranpose becomes matrix transposition CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 36

t-product Orthogonality Definition: An n n l real-valued tensor Q is orthogonal if Note that this means that Q Q = Q Q = I Q(:, i, :) Q(:, j, :) = { e1 i = j 0 i j CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 37

t-svd and Trunction Optimality Theorem: Let the T -SVD of A R l m n be given by A = U S V, with l l n orthogonal tensor U, m m n orthogonal tensor V, and l m n f-diagonal tensor S For k < min(l, m), define Then A k = U(:, 1 : k, :) S(1 : k, 1 : k, :) V (:, 1 : k, :) = A k = arg min A Â Â M where M = {C = X Y X R l k n, Y R k m n } k U(:, i, :) S(i, i, :) V(:, i, :) i=1 CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 38

t-svd and Optimality in Truncation Let A R m p n, for k < min(m, p), define A k = Then k U(:, i, :) S(i, i, :) V(:, i, :) i=1 A k = arg min A Ã Ã M where M = {C = X Y X R m k n, Y R k p n } CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 39

t-svd example Let A be 2 2 2 ] [Â1 (F I)circ (A) (F 0 I) = C 4 4 0 Â 2 [ ] ] [Â1 0 [Û1 0 = 0 Â 2 0 Û 2 ] ˆσ (1) 1 0 0 ˆσ (1) 2 [ ] ˆσ (2) 1 0 0 ˆσ (2) 2 [ ] ˆV 1 0 0 ˆV 2 The U, S, V T are formed by putting the hat matrices as frontal slices, then ifft along tubes [ ] ˆσ (1) e.g. S (1,1,:) obtained from ifft of vector 1 ˆσ (2) oriented into screen 1 CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 40

T -SVD and Multiway PCA X j, j = 1, 2,..., m are the training images M is the mean image A(:, j, :) = X j M stores the mean-subtracted images K = A A = U S S U is the covariance tensor Left orthogonal U contains the principal components with respect to K A(:, j, :) U(:, 1 : k, :) U(:, 1 : k, :) A(:, j, :) = k U(:, t, :) C(t, j, :) t=1 CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 41

T -QR Decomposition Theorem: Let A be an l m n real-valued tensor, then A can be factored as A P = Q R where Q is orthogonal l l n, R is l m n f-upper triangular, and P is a permutation tensor Cheaper for updating and downdating CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 42

Face Recognition Task Multilinear (Tensor) ICA and Dimensionality Reduction, M.A.O. Vasilescu, D. Terzopoulos, Proc. 7th International Conference on Independent Component Analysis and Signal Separation (ICA07), London, UK, September, 2007. In Lecture Notes in Computer Science, 4666, Springer-Verlag, New York, 2007, 818-826 CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 43

Face Recognition Task Experiment 1: randomly selected 15 images of each person as training set and test all remaining images Experiment 2: randomly selected 5 images of each person as the training set and test all remaining images Preprocessing: decimated the images by a factor of 3 to 64 56 pixels 20 trials for each experiment The Extended Yale Face Database B, http://vision.ucsd.edu/ leekc/extyaledatabase/extyaleb.html CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 44

T -SVD vs. PCA N. Hao, M.E. Kilmer, K. Braman, R.C. Hoover, Facial Recognition Using Tensor-Tensor Decompositions, SIAM J. Imaging Sci., 6(1), 437-463 CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 45

T -SVD vs. PCA N. Hao, M.E. Kilmer, K. Braman, R.C. Hoover, Facial Recognition Using Tensor-Tensor Decompositions, SIAM J. Imaging Sci., 6(1), 437-463 CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 46

T -QR vs. PCA N. Hao, M.E. Kilmer, K. Braman, R.C. Hoover, Facial Recognition Using Tensor-Tensor Decompositions, SIAM J. Imaging Sci., 6(1), 437-463 CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 47

T -QR vs. PCA N. Hao, M.E. Kilmer, K. Braman, R.C. Hoover, Facial Recognition Using Tensor-Tensor Decompositions, SIAM J. Imaging Sci., 6(1), 437-463 CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 48

Non-Negative Tensor Decompositions - t-product Given a nonnegative third-order tensor T R l m n and a positive integer k < min(l, m, n) Find nonnegative G R l k n, H R k m n such that min T G H 2 F Ĝ,Ĥ CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 49

Non-Negative Tensor Decompositions - t-product Given a nonnegative third-order tensor T R l m n and a positive integer k < min(l, m, n) Find nonnegative G R l k n, H R k m n such that min T G H 2 F Ĝ,Ĥ Facial Recognition Example: Dataset: The Center for Biological and Computational Learning (CBCL) Database Training images: 200 k = 10 CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 50

Reconstructed Images Based on NMF, NTF-CP and NTF-GH N. Hao, L. Horesh, M. Kilmer, Non-negative Tensor Decomposition, Compressed Sensing & Sparse Filtering, Springer, 123 148, 2014 CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 51

Tensor Nuclear Norm If A is an l m, l m matrix with singular values σ i, the nuclear norm A = m i=1 σ i. However, in the t-svd, we have singular tubes (the entries of which need not be positive), which sum up to a singular tube! The entries in the jth singular tube are the inverse Fourier coefficients of the length-n vector of the jth singular values of Â:,:,i, i = 1..n. Definition For A R l m n, our tensor nuclear norm is A = min(l,m) i=1 nf V s i 1 = min(l,m) n i=1 j=1 Ŝi,i,j. (Same as the matrix nuclear norm of circ (A)). CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 52

Tensor Nuclear Norm Theorem (Semerci,Hao,Kilmer,Miller) The tensor nuclear norm is a valid norm. Since the t-svd extends to higher-order tensors [Martin et al, 2012], the norm does, as well. CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 53

Tensor Completion Given unknown tensor T M of size n 1 n 2 n 3, given a subset of entries { T M ijk : (i, j, k) Ω} where Ω is an indicator tensor of size n 1 n 2 n 3. Recover the entire T M: min T X subject to P Ω ( T X) = P Ω ( T M) The (i, j, k) th component of P Ω ( T X) is equal to T M ijk if (i, j, k) Ω and zero otherwise. Similar to the previous problem, this can be solved by ADMM, with 3 update steps, one which decouples, one that is a shrinkage / thresholding step. CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 54

Numerical Results TNN minimization, Low Rank Tensor Completion (LRTC) [Liu, et al, 2013] based on tensor-n-rank [Gandy, et al, 2011], and the nuclear norm minimization on the vectorized video data [Cai, et al, 2010]. MERL 3 video, Basketball video 3 with thanks to A. Agrawal CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 55

Numerical Results CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 56

Toolbox and References Brett W. Bader, Tamara G. Kolda and others. MATLAB Tensor Toolbox Version 2.5, Available online, January 2012, http://www.sandia.gov/ tgkolda/tensortoolbox/ T. G. Kolda and B. W. Bader, Tensor Decompositions and Applications, SIAM Review 51(3):455-500, 2009 A. Cichocki, R. Zdunek, A.H. Phan, S.i. Amari, Nonnegative Matrix and Tensor Factorizations: Applications to Exploratory Multi-way Data Analysis and Blind Source Separation, 2009 M.E. Kilmer and C.D. Martin. Factorization strategies for third-order tensors, Linear Algebra and its Applications, Special Issue in Honor of G. W. Stewart s 70 th birthday, vol. 435(3):641 658, 2011 N. Hao, L. Horesh, M. Kilmer, Non-negative Tensor Decomposition, Compressed Sensing & Sparse Filtering, Springer, 123 148, 2014 CVPR 2017 New Tensor Algebra Lior Horesh & Misha Kilmer 57