A randomized block sampling approach to the canonical polyadic decomposition of large-scale tensors
|
|
- Egbert Ward
- 5 years ago
- Views:
Transcription
1 A randomized block sampling approach to the canonical polyadic decomposition of large-scale tensors Nico Vervliet Joint work with Lieven De Lathauwer SIAM AN17, July 13, 2017
2 2 Classification of hazardous gasses using e-noses Classify 900 experiments containing 72 time series with samples each. Sensor Experiment Time
3 3 Overview Decomposing large-scale tensors Randomized block sampling Experimental results Chemo-sensing application
4 4 Canonical polyadic decomposition Sum of R rank-1 terms c 1 c R T = a 1 b a R b R
5 4 Canonical polyadic decomposition Sum of R rank-1 terms c 1 c R T = a 1 b a R b R Mathematically, for a general Nth order tensor T T = R r=1 a (1) r a (2) r a (N) r = A (1), A (2),..., A (N)
6 5 Computing a CPD Optimization problem: 1 min A (1),A (2),...,A (N) 2 T A (1), A (2),..., A (N) 2 F
7 5 Computing a CPD Optimization problem: 1 min A (1),A (2),...,A (N) 2 T A (1), A (2),..., A (N) 2 Algorithms Alternating least squares CPOPT [Acar et al. 2011a] (Damped) Gauss Newton [Phan et al. 2013] (Inexact) nonlinear least squares [Sorber et al. 2013] F
8 6 Curse of dimensionality Suppose Nth order T C I I I, then number of entries: I N memory and time complexity: O ( I N)
9 6 Curse of dimensionality Suppose Nth order T C I I I, then number of entries: I N memory and time complexity: O ( I N) number of variables: NIR
10 6 Curse of dimensionality Suppose Nth order T C I I I, then number of entries: I N memory and time complexity: O ( I N) number of variables: NIR Example [Vervliet et al. 2014] Ninth-order tensor with I = 100 and rank R = 5: number of entries: number of variables: 4500
11 7 How to handle large tensors? Use incomplete tensors Acar et al. 2011b; Vervliet et al. 2014; Vervliet et al. 2016a Exploit sparsity Kang et al. 2012; Papalexakis et al. 2012; Bader and Kolda 2007 Compress the tensor Sidiropoulos et al. 2014; Oseledets and Tyrtyshnikov 2010; Vervliet et al. 2016b Decompose subtensors and combine results Papalexakis et al. 2012; Phan and Cichocki 2011 Parallel Liavas and Sidiropoulos many of the above
12 8 Overview Decomposing large-scale tensors Randomized block sampling Experimental results Chemo-sensing application
13 9 Randomized block sampling CPD: idea + +
14 9 Randomized block sampling CPD: idea + +
15 9 Randomized block sampling CPD: idea + + Take sample
16 9 Randomized block sampling CPD: idea + + Take sample Initialization Compute step + +
17 9 Randomized block sampling CPD: idea + + Take sample Initialization Update Compute step + +
18 10 Randomized block sampling CPD: algorithm input : Data T and initial guess A (n), n = 1,..., N output: A (n), n = 1,..., N such that T A (1),..., A (N) while k < K and not converged do Create sample T s and corresponding A (n) s, n = 1,..., N Let Ā (n) s be the result of 1 iteration in a restricted CPD algorithm on T s with initial guess A (n) s, n = 1,..., N and restriction Update the affected variables A (n) using Ā (n) s, n = 1,..., N k k + 1
19 10 Randomized block sampling CPD: algorithm input : Data T and initial guess A (n), n = 1,..., N output: A (n), n = 1,..., N such that T A (1),..., A (N) while k < K and not converged do Create sample T s and corresponding A (n) s, n = 1,..., N Let Ā (n) s be the result of 1 iteration in a restricted CPD algorithm on T s with initial guess A (n) s, n = 1,..., N and restriction Update the affected variables A (n) using Ā (n) s, n = 1,..., N k k + 1
20 11 Ingredient 1: randomized block sampling For a 6 6 tensor and block size 3 2: I 1 = {3, 1, 2, 6, 5, 4} I 2 = {1, 2, 4, 6, 3, 5}
21 11 Ingredient 1: randomized block sampling For a 6 6 tensor and block size 3 2: I 1 = {3, 1, 2, 6, 5, 4} I 2 = {1, 2, 4, 6, 3, 5} I 1 = {3, 1, 2, 6, 5, 4} I 2 = {1, 2, 4, 6, 3, 5}
22 11 Ingredient 1: randomized block sampling For a 6 6 tensor and block size 3 2: I 1 = {3, 1, 2, 6, 5, 4} I 2 = {1, 2, 4, 6, 3, 5} I 1 = {3, 1, 2, 6, 5, 4} I 2 = {1, 2, 4, 6, 3, 5} I 1 = {6, 1, 4, 2, 5, 3} I 2 = {1, 2, 4, 6, 3, 5}
23 12 Ingredient 2: restricted CPD algorithm ALS variant A (n) k+1 = (1 α)a(n) k + αt (n) V (n) ( W (n) ) 1 Enforce restriction by α = k.
24 12 Ingredient 2: restricted CPD algorithm ALS variant A (n) k+1 = (1 α)a(n) k + αt (n) V (n) ( W (n) ) 1 Enforce restriction by α = k. NLS variant in which 1 min p k 2 vec (F(x k)) J k p k 2 s.t. p k k F = T A (1),..., A (N)
25 13 Ingredient 3: restriction Use restriction of form { k = 0 ˆ 0 α (k Ksearch)/Q if if k < K search k K search Iteration k
26 13 Ingredient 3: restriction Use restriction of form { k = 0 ˆ 0 α (k Ksearch)/Q if if k < K search k K search Iteration k Example (Selecting Q) For a tensor and block size , Q = 4
27 14 Ingredient 4: A stopping criterion Function evaluation f val = 0.5 T A (1),..., A (N) f val CPD Error Iteration k
28 14 Ingredient 4: A stopping criterion Function evaluation f val = 0.5 T A (1),..., A (N) f val CPD Error Iteration k Step size
29 15 Intermezzo: Cramér Rao bound Uncertainty of an estimate 68% 3σ 2σ σ 0 σ 2σ 3σ
30 15 Intermezzo: Cramér Rao bound Uncertainty of an estimate 68% CRB σ 2 3σ 2σ σ 0 σ 2σ 3σ
31 15 Intermezzo: Cramér Rao bound Uncertainty of an estimate 68% CRB σ 2 3σ 2σ σ 0 σ 2σ 3σ C = τ 2 (J H J) 1
32 16 Ingredient 4: Cramér Rao bound based stopping criterion Experimental bound (n) Use estimates A k Use f val to estimate noise τ
33 16 Ingredient 4: Cramér Rao bound based stopping criterion Experimental bound (n) Use estimates A k Use f val to estimate noise τ Stopping criterion: 1 D CRB = R n I n N I n R n=1 i=1 r=1 A (n) k (i, r) A(n) k K CRB (i, r) C (n) (i, r)
34 16 Ingredient 4: Cramér Rao bound based stopping criterion Experimental bound (n) Use estimates A k Use f val to estimate noise τ Stopping criterion: 1 D CRB = R n I n γ N I n R n=1 i=1 r=1 A (n) k (i, r) A(n) k K CRB (i, r) C (n) (i, r)
35 17 Unrestricted phase vs restricted phase CPD Error Iteration k Unrestricted phase (1 + 2): converge to a neighborhood of an optimum Restricted phase (3): pull iterates towards optimum
36 17 Unrestricted phase vs restricted phase CPD Error Iteration k Unrestricted phase (1 + 2): converge to a neighborhood of an optimum Restricted phase (3): pull iterates towards optimum
37 17 Unrestricted phase vs restricted phase CPD Error Iteration k Unrestricted phase (1 + 2): converge to a neighborhood of an optimum Restricted phase (3): pull iterates towards optimum Assumptions CPD of rank R exists SNR is high enough Most block dimensions > R
38 18 Overview Decomposing large-scale tensors Randomized block sampling Experimental results Chemo-sensing application
39 19 Experiment overview Experiments Comparison ALS vs NLS (see paper) Influence of block size Influence of step size (see paper)
40 19 Experiment overview Experiments Comparison ALS vs NLS (see paper) Influence of block size Influence of step size (see paper) Performance 50 Monte Carlo experiments CPD error max A (n) n 0 A (n) res / A (n) 0
41 19 Experiment overview Experiments Comparison ALS vs NLS (see paper) Influence of block size Influence of step size (see paper) Performance 50 Monte Carlo experiments CPD error max A (n) n 0 A (n) res / A (n) 0 cpd rbs in Tensorlab 3.0 [Vervliet et al. 2016c]
42 20 Influence of block size: setup (4 4 2) ν U(0, 1) = N R = 20 No noise
43 21 Influence of block size on computation time 150 Time (s) full ν (4 4 2) ν R = 20, U(0, 1) No noise
44 22 Influence of block size on data accesses Data accesses (%) 1000 Full tensor full ν (4 4 2) ν R = 20, U(0, 1) No noise
45 23 Influence of block size on accuracy 10 0 Unrestricted ECPD full ν (4 4 2) ν R = 20, U(0, 1) 20 db
46 23 Influence of block size on accuracy ECPD Unrestricted Restricted full ν (4 4 2) ν R = 20, U(0, 1) 20 db
47 24 Overview Decomposing large-scale tensors Randomized block sampling Experimental results Chemo-sensing application
48 25 Classify hazardous gasses Does the sample contain CO, acetaldehyde or ammonia? Sensor Experiment Time Strategy: classify using coefficients of spatiotemporal patterns R = 5 Unknown
49 26 Classify hazardous gasses: results Resulting factor matrices time sensor experiment
50 26 Classify hazardous gasses: results Resulting factor matrices time sensor experiment Performance after clustering Iterations Time (s) Error (%) No restriction Restriction
51 27 Conclusion The randomized block sampling CPD algorithm enables the decomposition of larger tensors, using fewer data points and less memory Block size controls accuracy, data accesses and time Step size restriction improves accuracy Cramér Rao bound based stopping criterion combines noise and step information
52 28 More details: N. Vervliet and L. De Lathauwer [2016]. A Randomized Block Sampling Approach to Canonical Polyadic Decomposition of Large-Scale Tensors. In: IEEE Journal of Selected Topics in Signal Processing 10.2, pp
53 A randomized block sampling approach to the canonical polyadic decomposition of large-scale tensors Nico Vervliet Joint work with Lieven De Lathauwer SIAM AN17, July 13, 2017
54 2 References I Acar, E., D.M. Dunlavy, and T.G. Kolda (2011a). A scalable optimization approach for fitting canonical tensor decompositions. In: Journal of Chemometrics 25.2, pp Acar, E. et al. (2011b). Scalable tensor factorizations for incomplete data. In: Chemometrics and Intelligent Laboratory Systems 106.1, pp Bader, B.W. and T.G. Kolda (2007). Efficient MATLAB computations with sparse and factored tensors. In: SIAM J. Sci. Comput. 30.1, pp Kang, U. et al. (2012). GigaTensor: scaling tensor analysis up by 100 times-algorithms and discoveries. In: Proceedings of the 18th ACM SIGKDD international conference on Knowledge discovery and data mining. ACM, pp
55 3 References II Liavas, A. and N. Sidiropoulos (2015). Parallel Algorithms for Constrained Tensor Factorization via the Alternating Direction Method of Multipliers. In: IEEE Trans. Signal Process. PP.99, pp Oseledets, I.V. and E.E. Tyrtyshnikov (2010). TT-cross approximation for multidimensional arrays. In: Linear Algebra and its Applications 432.1, pp Papalexakis, E., C. Faloutsos, and N. Sidiropoulos (2012). ParCube: Sparse Parallelizable Tensor Decompositions. English. In: Machine Learning and Knowledge Discovery in Databases. Ed. by PeterA. Flach, Tijl De Bie, and Nello Cristianini. Vol Lecture Notes in Computer Science. Springer Berlin Heidelberg, pp
56 4 References III Phan, A.-H. and A. Cichocki (2011). PARAFAC algorithms for large-scale problems. In: Neurocomputing 74.11, pp Phan, A.-H., P. Tichavský, and A. Cichocki (2013). Low Complexity Damped Gauss Newton Algorithms for CANDECOMP/PARAFAC. In: SIAM J. Appl. Math. 34.1, pp Sidiropoulos, N., E. Papalexakis, and C. Faloutsos (2014). Parallel randomly compressed cubes: A scalable distributed architecture for big tensor decomposition. In: IEEE Signal Process. Mag. 31.5, pp
57 5 References IV Sorber, L., M. Van Barel, and L. De Lathauwer (2013). Optimization-Based Algorithms for Tensor Decompositions: Canonical Polyadic Decomposition, Decomposition in Rank-(L r, L r, 1) Terms, and a New Generalization. In: 23.2, pp Vervliet, N. and L. De Lathauwer (2016). A Randomized Block Sampling Approach to Canonical Polyadic Decomposition of Large-Scale Tensors. In: IEEE Journal of Selected Topics in Signal Processing 10.2, pp Vervliet, N., O. Debals, and L. De Lathauwer (2016a). Canonical polyadic decomposition of incomplete tensors with linearly constrained factors. Technical Report , ESAT-STADIUS, KU Leuven, Belgium.
58 6 References V (2016b). Tensorlab 3.0 Numerical optimization strategies for large-scale constrained and coupled matrix/tensor factorization. In: th Asilomar Conference on Signals, Systems and Computers. Vervliet, N. et al. (2014). Breaking the Curse of Dimensionality Using Decompositions of Incomplete Tensors: Tensor-based scientific computing in big data analysis. In: IEEE Signal Process. Mag. 31.5, pp Vervliet, N. et al. (2016c). Tensorlab 3.0. Available online at
Dealing with curse and blessing of dimensionality through tensor decompositions
Dealing with curse and blessing of dimensionality through tensor decompositions Lieven De Lathauwer Joint work with Nico Vervliet, Martijn Boussé and Otto Debals June 26, 2017 2 Overview Curse of dimensionality
More information/16/$ IEEE 1728
Extension of the Semi-Algebraic Framework for Approximate CP Decompositions via Simultaneous Matrix Diagonalization to the Efficient Calculation of Coupled CP Decompositions Kristina Naskovska and Martin
More informationCoupled Matrix/Tensor Decompositions:
Coupled Matrix/Tensor Decompositions: An Introduction Laurent Sorber Mikael Sørensen Marc Van Barel Lieven De Lathauwer KU Leuven Belgium Lieven.DeLathauwer@kuleuven-kulak.be 1 Canonical Polyadic Decomposition
More informationThe multiple-vector tensor-vector product
I TD MTVP C KU Leuven August 29, 2013 In collaboration with: N Vanbaelen, K Meerbergen, and R Vandebril Overview I TD MTVP C 1 Introduction Inspiring example Notation 2 Tensor decompositions The CP decomposition
More informationENGG5781 Matrix Analysis and Computations Lecture 10: Non-Negative Matrix Factorization and Tensor Decomposition
ENGG5781 Matrix Analysis and Computations Lecture 10: Non-Negative Matrix Factorization and Tensor Decomposition Wing-Kin (Ken) Ma 2017 2018 Term 2 Department of Electronic Engineering The Chinese University
More informationA variable projection method for block term decomposition of higher-order tensors
A variable projection method for block term decomposition of higher-order tensors Guillaume Olikier 1, P.-A. Absil 1, and Lieven De Lathauwer 2 1- Université catholique de Louvain - ICTEAM Institute B-1348
More informationDecompositions of Higher-Order Tensors: Concepts and Computation
L. De Lathauwer Decompositions of Higher-Order Tensors: Concepts and Computation Lieven De Lathauwer KU Leuven Belgium Lieven.DeLathauwer@kuleuven-kulak.be 1 L. De Lathauwer Canonical Polyadic Decomposition
More informationarxiv: v1 [cs.lg] 18 Nov 2018
THE CORE CONSISTENCY OF A COMPRESSED TENSOR Georgios Tsitsikas, Evangelos E. Papalexakis Dept. of Computer Science and Engineering University of California, Riverside arxiv:1811.7428v1 [cs.lg] 18 Nov 18
More informationLarge Scale Tensor Decompositions: Algorithmic Developments and Applications
Large Scale Tensor Decompositions: Algorithmic Developments and Applications Evangelos Papalexakis, U Kang, Christos Faloutsos, Nicholas Sidiropoulos, Abhay Harpale Carnegie Mellon University, KAIST, University
More informationComputing and decomposing tensors
Computing and decomposing tensors Tensor rank decomposition Nick Vannieuwenhoven (FWO / KU Leuven) Sensitivity Condition numbers Tensor rank decomposition Pencil-based algorithms Alternating least squares
More informationFitting a Tensor Decomposition is a Nonlinear Optimization Problem
Fitting a Tensor Decomposition is a Nonlinear Optimization Problem Evrim Acar, Daniel M. Dunlavy, and Tamara G. Kolda* Sandia National Laboratories Sandia is a multiprogram laboratory operated by Sandia
More informationModeling Parallel Wiener-Hammerstein Systems Using Tensor Decomposition of Volterra Kernels
Modeling Parallel Wiener-Hammerstein Systems Using Tensor Decomposition of Volterra Kernels Philippe Dreesen 1, David T. Westwick 2, Johan Schoukens 1, Mariya Ishteva 1 1 Vrije Universiteit Brussel (VUB),
More informationAvailable Ph.D position in Big data processing using sparse tensor representations
Available Ph.D position in Big data processing using sparse tensor representations Research area: advanced mathematical methods applied to big data processing. Keywords: compressed sensing, tensor models,
More informationPostgraduate Course Signal Processing for Big Data (MSc)
Postgraduate Course Signal Processing for Big Data (MSc) Jesús Gustavo Cuevas del Río E-mail: gustavo.cuevas@upm.es Work Phone: +34 91 549 57 00 Ext: 4039 Course Description Instructor Information Course
More informationFROM BASIS COMPONENTS TO COMPLEX STRUCTURAL PATTERNS Anh Huy Phan, Andrzej Cichocki, Petr Tichavský, Rafal Zdunek and Sidney Lehky
FROM BASIS COMPONENTS TO COMPLEX STRUCTURAL PATTERNS Anh Huy Phan, Andrzej Cichocki, Petr Tichavský, Rafal Zdunek and Sidney Lehky Brain Science Institute, RIKEN, Wakoshi, Japan Institute of Information
More informationThe Canonical Tensor Decomposition and Its Applications to Social Network Analysis
The Canonical Tensor Decomposition and Its Applications to Social Network Analysis Evrim Acar, Tamara G. Kolda and Daniel M. Dunlavy Sandia National Labs Sandia is a multiprogram laboratory operated by
More informationSparseness Constraints on Nonnegative Tensor Decomposition
Sparseness Constraints on Nonnegative Tensor Decomposition Na Li nali@clarksonedu Carmeliza Navasca cnavasca@clarksonedu Department of Mathematics Clarkson University Potsdam, New York 3699, USA Department
More informationScalable Tensor Factorizations with Incomplete Data
Scalable Tensor Factorizations with Incomplete Data Tamara G. Kolda & Daniel M. Dunlavy Sandia National Labs Evrim Acar Information Technologies Institute, TUBITAK-UEKAE, Turkey Morten Mørup Technical
More informationCP DECOMPOSITION AND ITS APPLICATION IN NOISE REDUCTION AND MULTIPLE SOURCES IDENTIFICATION
International Conference on Computer Science and Intelligent Communication (CSIC ) CP DECOMPOSITION AND ITS APPLICATION IN NOISE REDUCTION AND MULTIPLE SOURCES IDENTIFICATION Xuefeng LIU, Yuping FENG,
More informationMust-read Material : Multimedia Databases and Data Mining. Indexing - Detailed outline. Outline. Faloutsos
Must-read Material 15-826: Multimedia Databases and Data Mining Tamara G. Kolda and Brett W. Bader. Tensor decompositions and applications. Technical Report SAND2007-6702, Sandia National Laboratories,
More informationA FLEXIBLE MODELING FRAMEWORK FOR COUPLED MATRIX AND TENSOR FACTORIZATIONS
A FLEXIBLE MODELING FRAMEWORK FOR COUPLED MATRIX AND TENSOR FACTORIZATIONS Evrim Acar, Mathias Nilsson, Michael Saunders University of Copenhagen, Faculty of Science, Frederiksberg C, Denmark University
More informationMultiscale Tensor Decomposition
Multiscale Tensor Decomposition Alp Ozdemir 1, Mark A. Iwen 1,2 and Selin Aviyente 1 1 Department of Electrical and Computer Engineering, Michigan State University 2 Deparment of the Mathematics, Michigan
More informationthe tensor rank is equal tor. The vectorsf (r)
EXTENSION OF THE SEMI-ALGEBRAIC FRAMEWORK FOR APPROXIMATE CP DECOMPOSITIONS VIA NON-SYMMETRIC SIMULTANEOUS MATRIX DIAGONALIZATION Kristina Naskovska, Martin Haardt Ilmenau University of Technology Communications
More informationDISTRIBUTED LARGE-SCALE TENSOR DECOMPOSITION. s:
Author manuscript, published in "2014 IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), lorence : Italy (2014)" DISTRIBUTED LARGE-SCALE TENSOR DECOMPOSITION André L..
More informationFundamentals of Multilinear Subspace Learning
Chapter 3 Fundamentals of Multilinear Subspace Learning The previous chapter covered background materials on linear subspace learning. From this chapter on, we shall proceed to multiple dimensions with
More informationOptimization of Symmetric Tensor Computations
Optimization of Symmetric Tensor Computations Jonathon Cai Department of Computer Science, Yale University New Haven, CT 0650 Email: jonathon.cai@yale.edu Muthu Baskaran, Benoît Meister, Richard Lethin
More informationA PARALLEL ALGORITHM FOR BIG TENSOR DECOMPOSITION USING RANDOMLY COMPRESSED CUBES (PARACOMP)
A PARALLEL ALGORTHM FOR BG TENSOR DECOMPOSTON USNG RANDOMLY COMPRESSED CUBES PARACOMP N.D. Sidiropoulos Dept. of ECE, Univ. of Minnesota Minneapolis, MN 55455, USA E.E. Papalexakis, and C. Faloutsos Dept.
More informationNesterov-based Alternating Optimization for Nonnegative Tensor Completion: Algorithm and Parallel Implementation
Nesterov-based Alternating Optimization for Nonnegative Tensor Completion: Algorithm and Parallel Implementation Georgios Lourakis and Athanasios P. Liavas School of Electrical and Computer Engineering,
More informationShaden Smith * George Karypis. Nicholas D. Sidiropoulos. Kejun Huang * Abstract
Streaming Tensor Factorization for Infinite Data Sources Downloaded 0/5/8 to 60.94.64.33. Redistribution subject to SIAM license or copyright; see http://www.siam.org/journals/ojsa.php Shaden Smith * shaden.smith@intel.com
More informationStructured tensor missing-trace interpolation in the Hierarchical Tucker format Curt Da Silva and Felix J. Herrmann Sept. 26, 2013
Structured tensor missing-trace interpolation in the Hierarchical Tucker format Curt Da Silva and Felix J. Herrmann Sept. 6, 13 SLIM University of British Columbia Motivation 3D seismic experiments - 5D
More informationHigh Performance Parallel Tucker Decomposition of Sparse Tensors
High Performance Parallel Tucker Decomposition of Sparse Tensors Oguz Kaya INRIA and LIP, ENS Lyon, France SIAM PP 16, April 14, 2016, Paris, France Joint work with: Bora Uçar, CNRS and LIP, ENS Lyon,
More informationIntroduction to Tensors. 8 May 2014
Introduction to Tensors 8 May 2014 Introduction to Tensors What is a tensor? Basic Operations CP Decompositions and Tensor Rank Matricization and Computing the CP Dear Tullio,! I admire the elegance of
More informationNonnegative Tensor Factorization using a proximal algorithm: application to 3D fluorescence spectroscopy
Nonnegative Tensor Factorization using a proximal algorithm: application to 3D fluorescence spectroscopy Caroline Chaux Joint work with X. Vu, N. Thirion-Moreau and S. Maire (LSIS, Toulon) Aix-Marseille
More informationRecovering Tensor Data from Incomplete Measurement via Compressive Sampling
Recovering Tensor Data from Incomplete Measurement via Compressive Sampling Jason R. Holloway hollowjr@clarkson.edu Carmeliza Navasca cnavasca@clarkson.edu Department of Electrical Engineering Clarkson
More informationOnline Tensor Factorization for. Feature Selection in EEG
Online Tensor Factorization for Feature Selection in EEG Alric Althoff Honors Thesis, Department of Cognitive Science, University of California - San Diego Supervised by Dr. Virginia de Sa Abstract Tensor
More informationTENSOR APPROXIMATION TOOLS FREE OF THE CURSE OF DIMENSIONALITY
TENSOR APPROXIMATION TOOLS FREE OF THE CURSE OF DIMENSIONALITY Eugene Tyrtyshnikov Institute of Numerical Mathematics Russian Academy of Sciences (joint work with Ivan Oseledets) WHAT ARE TENSORS? Tensors
More informationThird-Order Tensor Decompositions and Their Application in Quantum Chemistry
Third-Order Tensor Decompositions and Their Application in Quantum Chemistry Tyler Ueltschi University of Puget SoundTacoma, Washington, USA tueltschi@pugetsound.edu April 14, 2014 1 Introduction A tensor
More informationDFacTo: Distributed Factorization of Tensors
DFacTo: Distributed Factorization of Tensors Joon Hee Choi Electrical and Computer Engineering Purdue University West Lafayette IN 47907 choi240@purdue.edu S. V. N. Vishwanathan Statistics and Computer
More informationModel-Driven Sparse CP Decomposition for Higher-Order Tensors
7 IEEE International Parallel and Distributed Processing Symposium Model-Driven Sparse CP Decomposition for Higher-Order Tensors Jiajia Li, Jee Choi, Ioakeim Perros, Jimeng Sun, Richard Vuduc Computational
More informationNonnegative Tensor Factorization with Smoothness Constraints
Nonnegative Tensor Factorization with Smoothness Constraints Rafal ZDUNEK 1 and Tomasz M. RUTKOWSKI 2 1 Institute of Telecommunications, Teleinformatics and Acoustics, Wroclaw University of Technology,
More informationDistributed Large-Scale Tensor Decomposition
Distributed Large-Scale Tensor Decomposition André L.. De Almeida, Alain Y. Kibangou To cite this version: André L.. De Almeida, Alain Y. Kibangou. Distributed Large-Scale Tensor Decomposition. 2014 IEEE
More informationBlind Source Separation of Single Channel Mixture Using Tensorization and Tensor Diagonalization
Blind Source Separation of Single Channel Mixture Using Tensorization and Tensor Diagonalization Anh-Huy Phan 1, Petr Tichavský 2(B), and Andrzej Cichocki 1,3 1 Lab for Advanced Brain Signal Processing,
More informationLocal Feature Extraction Models from Incomplete Data in Face Recognition Based on Nonnegative Matrix Factorization
American Journal of Software Engineering and Applications 2015; 4(3): 50-55 Published online May 12, 2015 (http://www.sciencepublishinggroup.com/j/ajsea) doi: 10.11648/j.ajsea.20150403.12 ISSN: 2327-2473
More informationarxiv: v1 [cs.lg] 3 Jul 2018
OCTen: Online Compression-based Tensor Decomposition Ekta Gujral UC Riverside egujr001@ucr.edu Ravdeep Pasricha UC Riverside rpasr001@ucr.edu Tianxiong Yang UC Riverside tyang022@ucr.edu Evangelos E. Papalexakis
More informationMulti-Way Compressed Sensing for Big Tensor Data
Multi-Way Compressed Sensing for Big Tensor Data Nikos Sidiropoulos Dept. ECE University of Minnesota MIIS, July 1, 2013 Nikos Sidiropoulos Dept. ECE University of Minnesota ()Multi-Way Compressed Sensing
More informationTime-Delay Estimation via CPD-GEVD Applied to Tensor-based GNSS Arrays with Errors
Time-Delay Estimation via CPD-GEVD Applied to Tensor-based GNSS Arrays with Errors Daniel Valle de Lima 1, João Paulo C. L. da Costa 1, Felix Antreich 2, Ricardo Kerhle Miranda 1, and Giovanni Del Galdo
More informationMATRIX COMPLETION AND TENSOR RANK
MATRIX COMPLETION AND TENSOR RANK HARM DERKSEN Abstract. In this paper, we show that the low rank matrix completion problem can be reduced to the problem of finding the rank of a certain tensor. arxiv:1302.2639v2
More informationARestricted Boltzmann machine (RBM) [1] is a probabilistic
1 Matrix Product Operator Restricted Boltzmann Machines Cong Chen, Kim Batselier, Ching-Yun Ko, and Ngai Wong chencong@eee.hku.hk, k.batselier@tudelft.nl, cyko@eee.hku.hk, nwong@eee.hku.hk arxiv:1811.04608v1
More informationMath 671: Tensor Train decomposition methods
Math 671: Eduardo Corona 1 1 University of Michigan at Ann Arbor December 8, 2016 Table of Contents 1 Preliminaries and goal 2 Unfolding matrices for tensorized arrays The Tensor Train decomposition 3
More informationEfficient CP-ALS and Reconstruction From CP
Efficient CP-ALS and Reconstruction From CP Jed A. Duersch & Tamara G. Kolda Sandia National Laboratories Livermore, CA Sandia National Laboratories is a multimission laboratory managed and operated by
More informationA Block-Jacobi Algorithm for Non-Symmetric Joint Diagonalization of Matrices
A Block-Jacobi Algorithm for Non-Symmetric Joint Diagonalization of Matrices ao Shen and Martin Kleinsteuber Department of Electrical and Computer Engineering Technische Universität München, Germany {hao.shen,kleinsteuber}@tum.de
More informationOrthogonal tensor decomposition
Orthogonal tensor decomposition Daniel Hsu Columbia University Largely based on 2012 arxiv report Tensor decompositions for learning latent variable models, with Anandkumar, Ge, Kakade, and Telgarsky.
More informationarxiv: v2 [cs.lg] 9 May 2018
TensorLy: Tensor Learning in Python arxiv:1610.09555v2 [cs.lg] 9 May 2018 Jean Kossaifi 1 jean.kossaifi@imperial.ac.uk Yannis Panagakis 1,2 i.panagakis@imperial.ac.uk Anima Anandkumar 3,4 anima@amazon.com
More informationarxiv: v4 [math.na] 10 Nov 2014
NEWTON-BASED OPTIMIZATION FOR KULLBACK-LEIBLER NONNEGATIVE TENSOR FACTORIZATIONS SAMANTHA HANSEN, TODD PLANTENGA, TAMARA G. KOLDA arxiv:134.4964v4 [math.na] 1 Nov 214 Abstract. Tensor factorizations with
More informationAn Effective Tensor Completion Method Based on Multi-linear Tensor Ring Decomposition
An Effective Tensor Completion Method Based on Multi-linear Tensor Ring Decomposition Jinshi Yu, Guoxu Zhou, Qibin Zhao and Kan Xie School of Automation, Guangdong University of Technology, Guangzhou,
More informationKronecker Product Approximation with Multiple Factor Matrices via the Tensor Product Algorithm
Kronecker Product Approximation with Multiple actor Matrices via the Tensor Product Algorithm King Keung Wu, Yeung Yam, Helen Meng and Mehran Mesbahi Department of Mechanical and Automation Engineering,
More informationBig Tensor Data Reduction
Big Tensor Data Reduction Nikos Sidiropoulos Dept. ECE University of Minnesota NSF/ECCS Big Data, 3/21/2013 Nikos Sidiropoulos Dept. ECE University of Minnesota () Big Tensor Data Reduction NSF/ECCS Big
More informationA BLIND SPARSE APPROACH FOR ESTIMATING CONSTRAINT MATRICES IN PARALIND DATA MODELS
2th European Signal Processing Conference (EUSIPCO 22) Bucharest, Romania, August 27-3, 22 A BLIND SPARSE APPROACH FOR ESTIMATING CONSTRAINT MATRICES IN PARALIND DATA MODELS F. Caland,2, S. Miron 2 LIMOS,
More informationNumerical Methods. Rafał Zdunek Underdetermined problems (2h.) Applications) (FOCUSS, M-FOCUSS,
Numerical Methods Rafał Zdunek Underdetermined problems (h.) (FOCUSS, M-FOCUSS, M Applications) Introduction Solutions to underdetermined linear systems, Morphological constraints, FOCUSS algorithm, M-FOCUSS
More informationTensor MUSIC in Multidimensional Sparse Arrays
Tensor MUSIC in Multidimensional Sparse Arrays Chun-Lin Liu 1 and P. P. Vaidyanathan 2 Dept. of Electrical Engineering, MC 136-93 California Institute of Technology, Pasadena, CA 91125, USA cl.liu@caltech.edu
More informationDimitri Nion & Lieven De Lathauwer
he decomposition of a third-order tensor in block-terms of rank-l,l, Model, lgorithms, Uniqueness, Estimation of and L Dimitri Nion & Lieven De Lathauwer.U. Leuven, ortrijk campus, Belgium E-mails: Dimitri.Nion@kuleuven-kortrijk.be
More informationAll-at-once Decomposition of Coupled Billion-scale Tensors in Apache Spark
All-at-once Decomposition of Coupled Billion-scale Tensors in Apache Spark Aditya Gudibanda, Tom Henretty, Muthu Baskaran, James Ezick, Richard Lethin Reservoir Labs 632 Broadway Suite 803 New York, NY
More informationA Practical Randomized CP Tensor Decomposition
A Practical Randomized CP Tensor Decomposition Casey Battaglino, Grey Ballard 2, and Tamara G. Kolda 3 SIAM AN 207, Pittsburgh, PA Georgia Tech Computational Sci. and Engr. 2 Wake Forest University 3 Sandia
More informationA Medium-Grained Algorithm for Distributed Sparse Tensor Factorization
A Medium-Grained Algorithm for Distributed Sparse Tensor Factorization Shaden Smith, George Karypis Department of Computer Science and Engineering, University of Minnesota {shaden, karypis}@cs.umn.edu
More informationA Simpler Approach to Low-Rank Tensor Canonical Polyadic Decomposition
A Simpler Approach to Low-ank Tensor Canonical Polyadic Decomposition Daniel L. Pimentel-Alarcón University of Wisconsin-Madison Abstract In this paper we present a simple and efficient method to compute
More informationMULTIPLICATIVE ALGORITHM FOR CORRENTROPY-BASED NONNEGATIVE MATRIX FACTORIZATION
MULTIPLICATIVE ALGORITHM FOR CORRENTROPY-BASED NONNEGATIVE MATRIX FACTORIZATION Ehsan Hosseini Asl 1, Jacek M. Zurada 1,2 1 Department of Electrical and Computer Engineering University of Louisville, Louisville,
More informationWindow-based Tensor Analysis on High-dimensional and Multi-aspect Streams
Window-based Tensor Analysis on High-dimensional and Multi-aspect Streams Jimeng Sun Spiros Papadimitriou Philip S. Yu Carnegie Mellon University Pittsburgh, PA, USA IBM T.J. Watson Research Center Hawthorne,
More informationCVPR A New Tensor Algebra - Tutorial. July 26, 2017
CVPR 2017 A New Tensor Algebra - Tutorial Lior Horesh lhoresh@us.ibm.com Misha Kilmer misha.kilmer@tufts.edu July 26, 2017 Outline Motivation Background and notation New t-product and associated algebraic
More informationTENLAB A MATLAB Ripoff for Tensors
TENLAB A MATLAB Ripoff for Tensors Y. Cem Sübakan, ys2939 Mehmet K. Turkcan, mkt2126 Dallas Randal Jones, drj2115 February 9, 2016 Introduction MATLAB is a great language for manipulating arrays. However,
More informationParallel Numerical Algorithms
Parallel Numerical Algorithms Chapter 6 Structured and Low Rank Matrices Section 6.3 Numerical Optimization Michael T. Heath and Edgar Solomonik Department of Computer Science University of Illinois at
More informationA Randomized Approach for Crowdsourcing in the Presence of Multiple Views
A Randomized Approach for Crowdsourcing in the Presence of Multiple Views Presenter: Yao Zhou joint work with: Jingrui He - 1 - Roadmap Motivation Proposed framework: M2VW Experimental results Conclusion
More informationNovel Alternating Least Squares Algorithm for Nonnegative Matrix and Tensor Factorizations
Novel Alternating Least Squares Algorithm for Nonnegative Matrix and Tensor Factorizations Anh Huy Phan 1, Andrzej Cichocki 1,, Rafal Zdunek 1,2,andThanhVuDinh 3 1 Lab for Advanced Brain Signal Processing,
More informationAn Introduction to Hierachical (H ) Rank and TT Rank of Tensors with Examples
An Introduction to Hierachical (H ) Rank and TT Rank of Tensors with Examples Lars Grasedyck and Wolfgang Hackbusch Bericht Nr. 329 August 2011 Key words: MSC: hierarchical Tucker tensor rank tensor approximation
More informationAnomaly Detection in Temporal Graph Data: An Iterative Tensor Decomposition and Masking Approach
Proceedings 1st International Workshop on Advanced Analytics and Learning on Temporal Data AALTD 2015 Anomaly Detection in Temporal Graph Data: An Iterative Tensor Decomposition and Masking Approach Anna
More informationParCube: Sparse Parallelizable Tensor Decompositions
ParCube: Sparse Parallelizable Tensor Decompositions Evangelos E. Papalexakis, Christos Faloutsos, and Nicholas D. Sidiropoulos 2 School of Computer Science, Carnegie Mellon University, Pittsburgh, PA,
More informationarxiv: v1 [math.ra] 13 Jan 2009
A CONCISE PROOF OF KRUSKAL S THEOREM ON TENSOR DECOMPOSITION arxiv:0901.1796v1 [math.ra] 13 Jan 2009 JOHN A. RHODES Abstract. A theorem of J. Kruskal from 1977, motivated by a latent-class statistical
More informationIntroduction to the Tensor Train Decomposition and Its Applications in Machine Learning
Introduction to the Tensor Train Decomposition and Its Applications in Machine Learning Anton Rodomanov Higher School of Economics, Russia Bayesian methods research group (http://bayesgroup.ru) 14 March
More informationUncorrelated Multilinear Principal Component Analysis through Successive Variance Maximization
Uncorrelated Multilinear Principal Component Analysis through Successive Variance Maximization Haiping Lu 1 K. N. Plataniotis 1 A. N. Venetsanopoulos 1,2 1 Department of Electrical & Computer Engineering,
More informationA concise proof of Kruskal s theorem on tensor decomposition
A concise proof of Kruskal s theorem on tensor decomposition John A. Rhodes 1 Department of Mathematics and Statistics University of Alaska Fairbanks PO Box 756660 Fairbanks, AK 99775 Abstract A theorem
More informationCS60021: Scalable Data Mining. Dimensionality Reduction
J. Leskovec, A. Rajaraman, J. Ullman: Mining of Massive Datasets, http://www.mmds.org 1 CS60021: Scalable Data Mining Dimensionality Reduction Sourangshu Bhattacharya Assumption: Data lies on or near a
More informationc 2008 Society for Industrial and Applied Mathematics
SIAM J MATRIX ANAL APPL Vol 30, No 3, pp 1219 1232 c 2008 Society for Industrial and Applied Mathematics A JACOBI-TYPE METHOD FOR COMPUTING ORTHOGONAL TENSOR DECOMPOSITIONS CARLA D MORAVITZ MARTIN AND
More informationTHE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR
THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR WEN LI AND MICHAEL K. NG Abstract. In this paper, we study the perturbation bound for the spectral radius of an m th - order n-dimensional
More informationA new truncation strategy for the higher-order singular value decomposition
A new truncation strategy for the higher-order singular value decomposition Nick Vannieuwenhoven K.U.Leuven, Belgium Workshop on Matrix Equations and Tensor Techniques RWTH Aachen, Germany November 21,
More informationTensor Decompositions for Signal Processing Applications
Tensor Decompositions for Signal Processing Applications From Two-way to Multiway Component Analysis A. Cichocki, D. Mandic, A-H. Phan, C. Caiafa, G. Zhou, Q. Zhao, and L. De Lathauwer Summary The widespread
More informationTechnical Report TR SPLATT: Efficient and Parallel Sparse Tensor-Matrix Multiplication
Technical Report Department of Computer Science and Engineering University of Minnesota 4-192 Keller Hall 200 Union Street SE Minneapolis, MN 55455-0159 USA TR 15-008 SPLATT: Efficient and Parallel Sparse
More informationComputational Linear Algebra
Computational Linear Algebra PD Dr. rer. nat. habil. Ralf-Peter Mundani Computation in Engineering / BGU Scientific Computing in Computer Science / INF Winter Term 2018/19 Part 6: Some Other Stuff PD Dr.
More informationFaloutsos, Tong ICDE, 2009
Large Graph Mining: Patterns, Tools and Case Studies Christos Faloutsos Hanghang Tong CMU Copyright: Faloutsos, Tong (29) 2-1 Outline Part 1: Patterns Part 2: Matrix and Tensor Tools Part 3: Proximity
More information3D INTERPOLATION USING HANKEL TENSOR COMPLETION BY ORTHOGONAL MATCHING PURSUIT A. Adamo, P. Mazzucchelli Aresys, Milano, Italy
3D INTERPOLATION USING HANKEL TENSOR COMPLETION BY ORTHOGONAL MATCHING PURSUIT A. Adamo, P. Mazzucchelli Aresys, Milano, Italy Introduction. Seismic data are often sparsely or irregularly sampled along
More informationPermutation transformations of tensors with an application
DOI 10.1186/s40064-016-3720-1 RESEARCH Open Access Permutation transformations of tensors with an application Yao Tang Li *, Zheng Bo Li, Qi Long Liu and Qiong Liu *Correspondence: liyaotang@ynu.edu.cn
More informationFrom Stationary Methods to Krylov Subspaces
Week 6: Wednesday, Mar 7 From Stationary Methods to Krylov Subspaces Last time, we discussed stationary methods for the iterative solution of linear systems of equations, which can generally be written
More informationUsing Hankel structured low-rank approximation for sparse signal recovery
Using Hankel structured low-rank approximation for sparse signal recovery Ivan Markovsky 1 and Pier Luigi Dragotti 2 Department ELEC Vrije Universiteit Brussel (VUB) Pleinlaan 2, Building K, B-1050 Brussels,
More informationTruncation Strategy of Tensor Compressive Sensing for Noisy Video Sequences
Journal of Information Hiding and Multimedia Signal Processing c 2016 ISSN 207-4212 Ubiquitous International Volume 7, Number 5, September 2016 Truncation Strategy of Tensor Compressive Sensing for Noisy
More informationBlind Parallel Interrogation of Ultrasonic Neural Dust Motes Based on Canonical Polyadic Decomposition: a Simulation Study
Blind Parallel Interrogation of Ultrasonic Neural Dust Motes Based on Canonical Polyadic Decomposition: a Simulation Study Alexander Bertrand, Dongjin Seo, Jose M. Carmena, Michel M. Maharbiz, Elad Alon
More informationFuture Directions in Tensor-Based Computation and Modeling 1
Workshop on Future Directions in Tensor-Based Computation and Modeling 1 National Science Foundation Arlington, Virginia February 20-21, 2009 Program Organizer: Charles Van Loan (Department of Computer
More informationInstitute for Computational Mathematics Hong Kong Baptist University
Institute for Computational Mathematics Hong Kong Baptist University ICM Research Report 08-0 How to find a good submatrix S. A. Goreinov, I. V. Oseledets, D. V. Savostyanov, E. E. Tyrtyshnikov, N. L.
More informationJOS M.F. TEN BERGE SIMPLICITY AND TYPICAL RANK RESULTS FOR THREE-WAY ARRAYS
PSYCHOMETRIKA VOL. 76, NO. 1, 3 12 JANUARY 2011 DOI: 10.1007/S11336-010-9193-1 SIMPLICITY AND TYPICAL RANK RESULTS FOR THREE-WAY ARRAYS JOS M.F. TEN BERGE UNIVERSITY OF GRONINGEN Matrices can be diagonalized
More informationTurbo-SMT: Accelerating Coupled Sparse Matrix-Tensor Factorizations by 200x
Turbo-SMT: Accelerating Coupled Sparse Matrix-Tensor Factorizations by 2x Evangelos E. Papalexakis epapalex@cs.cmu.edu Christos Faloutsos christos@cs.cmu.edu Tom M. Mitchell tom.mitchell@cmu.edu Partha
More informationSelected Topics in Optimization. Some slides borrowed from
Selected Topics in Optimization Some slides borrowed from http://www.stat.cmu.edu/~ryantibs/convexopt/ Overview Optimization problems are almost everywhere in statistics and machine learning. Input Model
More informationA Brief Guide for TDALAB Ver 1.1. Guoxu Zhou and Andrzej Cichocki
A Brief Guide for TDALAB Ver 1.1 Guoxu Zhou and Andrzej Cichocki April 30, 2013 Contents 1 Preliminary 2 1.1 Highlights of TDALAB...................... 2 1.2 Install and Run TDALAB....................
More informationMath 671: Tensor Train decomposition methods II
Math 671: Tensor Train decomposition methods II Eduardo Corona 1 1 University of Michigan at Ann Arbor December 13, 2016 Table of Contents 1 What we ve talked about so far: 2 The Tensor Train decomposition
More information