Tensor Decompositions and Applications

Similar documents
Introduction to Tensors. 8 May 2014

Computational Linear Algebra

Fitting a Tensor Decomposition is a Nonlinear Optimization Problem

Parallel Tensor Compression for Large-Scale Scientific Data

Lecture 4. CP and KSVD Representations. Charles F. Van Loan

Tensor Decompositions for Machine Learning. G. Roeder 1. UBC Machine Learning Reading Group, June University of British Columbia

Efficient CP-ALS and Reconstruction From CP

Faloutsos, Tong ICDE, 2009

This work has been submitted to ChesterRep the University of Chester s online research repository.

1. Connecting to Matrix Computations. Charles F. Van Loan

Tensor Analysis. Topics in Data Mining Fall Bruno Ribeiro

The Canonical Tensor Decomposition and Its Applications to Social Network Analysis

From Matrix to Tensor. Charles F. Van Loan

SPARSE TENSORS DECOMPOSITION SOFTWARE

Lecture 4. Tensor-Related Singular Value Decompositions. Charles F. Van Loan

CP DECOMPOSITION AND ITS APPLICATION IN NOISE REDUCTION AND MULTIPLE SOURCES IDENTIFICATION

Scalable Tensor Factorizations with Incomplete Data

Lecture 2. Tensor Unfoldings. Charles F. Van Loan

CVPR A New Tensor Algebra - Tutorial. July 26, 2017

/16/$ IEEE 1728

DFacTo: Distributed Factorization of Tensors

Fundamentals of Multilinear Subspace Learning

Optimization of Symmetric Tensor Computations

Math 671: Tensor Train decomposition methods

Third-Order Tensor Decompositions and Their Application in Quantum Chemistry

A Practical Randomized CP Tensor Decomposition

Coupled Matrix/Tensor Decompositions:

Multiway Analysis of Bridge Structural Types in the National Bridge Inventory (NBI) A Tensor Decomposition Approach

Big Data Analytics. Special Topics for Computer Science CSE CSE Jan 21

Tutorial on MATLAB for tensors and the Tucker decomposition

BOOLEAN TENSOR FACTORIZATIONS. Pauli Miettinen 14 December 2011

Sparseness Constraints on Nonnegative Tensor Decomposition

The multiple-vector tensor-vector product

FACTORIZATION STRATEGIES FOR THIRD-ORDER TENSORS

Kronecker Product Approximation with Multiple Factor Matrices via the Tensor Product Algorithm

CS 143 Linear Algebra Review

UNIT 6: The singular value decomposition.

Large Scale Data Analysis Using Deep Learning

Math 671: Tensor Train decomposition methods II

Numerical Methods I Non-Square and Sparse Linear Systems

Clustering Boolean Tensors

Strassen-like algorithms for symmetric tensor contractions

Lecture 1: Introduction to low-rank tensor representation/approximation. Center for Uncertainty Quantification. Alexander Litvinenko

Tensor Network Computations in Quantum Chemistry. Charles F. Van Loan Department of Computer Science Cornell University

BlockMatrixComputations and the Singular Value Decomposition. ATaleofTwoIdeas

Illustration by Chris Brigman IPAM Workshop: Big Data Meets Computation February 2, 2017

The study of linear algebra involves several types of mathematical objects:

Illustration by Chris Brigman Michigan Institute for Data Science (MIDAS) Seminar November 9, 2015

PAVEMENT CRACK CLASSIFICATION BASED ON TENSOR FACTORIZATION. Offei Amanor Adarkwa

Computing Dense Tensor Decompositions with Optimal Dimension Trees

Must-read Material : Multimedia Databases and Data Mining. Indexing - Detailed outline. Outline. Faloutsos

Multi-Linear Mappings, SVD, HOSVD, and the Numerical Solution of Ill-Conditioned Tensor Least Squares Problems

the tensor rank is equal tor. The vectorsf (r)

An Introduction to Hierachical (H ) Rank and TT Rank of Tensors with Examples

Tensor Decompositions Workshop Discussion Notes American Institute of Mathematics (AIM) Palo Alto, CA

Matrix-Product-States/ Tensor-Trains

Linear Algebra for Machine Learning. Sargur N. Srihari

arxiv: v4 [math.na] 10 Nov 2014

Introduction to the Tensor Train Decomposition and Its Applications in Machine Learning

Linear Algebraic Equations

Algorithm 862: MATLAB Tensor Classes for Fast Algorithm Prototyping

On the numerical solution of the chemical master equation with sums of rank one tensors

Inequalities Involving Khatri-Rao Products of Positive Semi-definite Matrices

Numerical tensor methods and their applications

1. Structured representation of high-order tensors revisited. 2. Multi-linear algebra (MLA) with Kronecker-product data.

The Singular Value Decomposition and Least Squares Problems

Discovering Fast Matrix Multiplication Algorithms via Tensor Decomposition

An Investigation of Sparse Tensor Formats for Tensor Libraries. Parker Allen Tew

Window-based Tensor Analysis on High-dimensional and Multi-aspect Streams

Bindel, Fall 2009 Matrix Computations (CS 6210) Week 8: Friday, Oct 17

Recovering Tensor Data from Incomplete Measurement via Compressive Sampling

Diagonal and Monomial Solutions of the Matrix Equation AXB = C

Tensor-Tensor Product Toolbox

Lecture 15 Review of Matrix Theory III. Dr. Radhakant Padhi Asst. Professor Dept. of Aerospace Engineering Indian Institute of Science - Bangalore

Nonnegative Tensor Factorization using a proximal algorithm: application to 3D fluorescence spectroscopy

AM205: Assignment 2. i=1

Deep Learning Book Notes Chapter 2: Linear Algebra

A Practical Randomized CP Tensor Decomposition

THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR

Vector and Matrix Norms. Vector and Matrix Norms

Orthogonal tensor decomposition

Non-negative Tensor Factorization Based on Alternating Large-scale Non-negativity-constrained Least Squares

Compressive Sensing of Sparse Tensor

Statistical Performance of Convex Tensor Decomposition

Towards a Regression using Tensors

linearly indepedent eigenvectors as the multiplicity of the root, but in general there may be no more than one. For further discussion, assume matrice

Data Mining and Matrices

Cartesian Tensors. e 2. e 1. General vector (formal definition to follow) denoted by components

Multi-Way Compressed Sensing for Big Tensor Data

Data Mining Lecture 4: Covariance, EVD, PCA & SVD

ENGG5781 Matrix Analysis and Computations Lecture 10: Non-Negative Matrix Factorization and Tensor Decomposition

Multilinear Singular Value Decomposition for Two Qubits

A new truncation strategy for the higher-order singular value decomposition

Enroatrso.o ay IEIR' Elmer" decomposition. I= ±Erob roer. decomposition. IE,a YoEYo. The CP decomposition. decomposition - given.

STA141C: Big Data & High Performance Statistical Computing

Tensor Decomposition for Fast Parsing with Latent-Variable PCFGs

Lecture 2: Linear Algebra Review

On the Simulatability Condition in Key Generation Over a Non-authenticated Public Channel

Large Scale Tensor Decompositions: Algorithmic Developments and Applications

New algorithms for tensor decomposition based on a reduced functional

Transcription:

Tamara G. Kolda and Brett W. Bader Part I September 22, 2015

What is tensor? A N-th order tensor is an element of the tensor product of N vector spaces, each of which has its own coordinate system. a = a i e i b = b j ẽ j c = c k ê k (Vector spaces) A = a b = a i b j = Aij e i ẽ j (Matrix, second order tensor) X = a b c = a i b j c k e i ẽ j ê k (Third order tensor) = X ijk... N-th order (ways or modes) tensor has N dimensions.

From matrix to high order tensor Table : Matrix to high order tensor Matrix High order tensor (Columns/Rows)A i, A j (Fibers)X ij, X i k, X jk ; (Slices) X i, X j, X k mode-3, 2, 1; Horizontal, Lateral, Frontal slices A B = A ij B ij X Y = X ijk Y ijk i=1 j=1 i=1 j=1 k=1 A F = A ij X = X i=1 j=1 (Rank one matrix) A = ab T A ij = a i b j (Symmetric) A = A T A = I, A ij = δ ij ijk i=1 j=1 k=1 (Rank one tensor) X = a b c X ijk = a i b j c k (Supersymmetric: cubical + symmetry) X R I I I X ijk is constant when permuting i, j, k X = I, X ijk = δ ijk

Matricization (unfolding, flattening) Vectorization of Matrix: arrange all columns of the matrix to be a column vector. ( 1 3 1 2 4 ) 2 3 4 Mode-n matricization of a tensor: arrange all mode-n fibers to be the columns of the resulting matrix. ( Let s skip the formal definition. ) 12 4 7 X ij1 = 2 10 6 1 16 9 13 3 73 X ij2 = 21 0 26 11 27 19 Mode-1 matricization: X (1) 12 4 7 13 3 73 2 10 6 21 0 26 1 16 9 11 27 19

Mode-2 matricization: X (2) 12 2 1 13 21 11 4 10 16 3 0 27 7 6 9 73 26 19 Mode-3 matricization: X (3) 12 2 1 4 10 16 6 9 ( 13 21 11 3 0 27 26 19 ) How do we change the results if we have an additional slice? 1 1 1 X ij3 = 1 1 1 1 1 1

Tensor multiplication: n-mode product The n-mode product of a tensor X R I 1 R I 2 R I N with a matrix U R J In is denoted by X n U. (X n U) i1 i n 1 ji n+1 i N = x i1 i ni n+1 i N u jin i n=1 I n Y = X n U Y (n) = UX (n) related to change of basis, when a tensor defines a multilinear operator. The n-mode product of a tensor X R I 1 R I 2 R I N with a vector v R In is denoted by X nv. I n (X nv) i1 i n 1 i n+1 i N = x i1 i N v in i n=1 precedence matters, contraction of tensor.

Matrix Kronecker,Khatri Rao and Hadamard Products A, B a 11 B a 1J B Kronecker product: A B = a I 1 B a IJ B Khatri Rao product: A B = [a 1 b 1 Hadamard Product: (A B) ij = A ij B ij Examples: MATLAB: kron(a,b); % Kronecker for i = 1:3 % Khatri-Rao C(:,i) = kron(a(:,i),b(:,i)); end A.*B %Hadamard 1 2 3 A = 4 5 6 7 8 9 a 2 b 2 a K b K ] (column matching) B = ( 1 2 3 4 5 6 ) A B?A B?A B?A A?

Tensor Rank and the CP Decomposition CP Decomposition (CANDECOMP/PARAFAC Decomposition), factorizes a tensor into a sum of component rank-one tensors. CP decomposition can be treated as a generalization of SVD to higher order tensors. R A = σ r u r v r, with σ 1 σ 2 σ R > 0 (SVD) r=1 X R I J K R λ r a r b r c r = λ; A, B, C r=1 A, B, and C are combination of the vectors from the rank-one components and are normalized to length one. In matricized form: X (1) AΛ(C B) T, X (2) BΛ(C A) T, X (3) CΛ(B A) T where Λ = diag(λ). It can be extended to n-th order tensor.

Tensor Rank The rank of a tensor X, denoted rank(x ), is the smallest number of rank-one tensors that generate X as their sum. (the smallest number of components in an exact = CP decomposition. In other words, what is R?) An exact CP decomposition with R = rank(x ) components is called the rank decomposition. Definition is an exact analogue from matrix rank (the dimension of the vector space spanned by its columns). Properties are quite different: (1) Field dependent: may not the same over R and C. (2) No straightforward algorithm to compute the rank of specific tensor (NP-hard problem). Matrix decompositions are not unique (why?), however high order tensor decompositions are often unique.

Low-Rank Approximations Matrix: a best rank-k approximation is given by the leading k factors of the SVD. A rank-k approximation that minimizes A B is given by: k A = σ r u r v r, with σ 1 σ 2 σ R > 0 r=1 Not true for high order tensors! a best rank-k approximation may not exist! (degeneracy)

Computing CP Decomposition Suppose R is fixed, for a third order tensor X, we are looking for CP Decomposition of ˆX : min X ˆX with ˆX = λ; A, B, C ˆX The alternating least squares (ALS) method: (fixed all but one matrix) Fix C and B (specify in any way) and rewrite (how?), Â = AΛ min X (1) Â(C B) T F Â Solution: Â = X (1) [(C B) T ], or a better version: Â = X (1) (C B)[(C T C B T B)]. refers to Moore Penrose pseudoinverse (MATLAB: B = pinv(a)). Normalization to get A, storing norms as λ. Fix A and B to compute C., until reach convergence (i.e. error stops decreasing). Implementability? Convergence? Efficiency?

Tensor toolbox MATLAB Sandia National Laboratories provide a MATLAB Tensor Toolbox which includes an implementation of CP Decomposition. cp als function implements ALS algorithm for CP Decomposition with fixed rank. e.g P = cp als(x, 2); other approaches are avaliable cp apr cp nmu, cp opt, etc. special treatment for sparse tensor: a greedy CP. Applications: signal/image processing neuroscience data mining (user keyword time: chatroom tensors) stochastic PDEs.

THANK YOU!