Statistical Performance of Convex Tensor Decomposition

Size: px
Start display at page:

Download "Statistical Performance of Convex Tensor Decomposition"

Transcription

1 Slides available: h-p:// tokyo.ac.jp/ryotat/tensor12kyoto.pdf Statistical Performance of Convex Tensor Decomposition Ryota Tomioka Kyoto University Perspectives in Informatics 4B Collaborators: Taiji Suzuki Kohei Hayashi Hisashi Kashima

2 Netflix challenge ( ) $1,000,000 prize Goal: Improve the performance of a video recommendafon system (predict who likes which movies) Example: Likes Star Wars and E.T., Doesn t like Minority Report. Does he like Blade Runner?

3 Matrix completion view User A User B User C User D ?? +1?? +1?? +1 1? +1 +1??? Goal: fill the missing entries!

4 Matrix completion Impossible without an assumpfon. (Missing entries can be arbitrary) problem is ill posed Most common assumpfon: Low rank decomposifon Y U V T

5 Matrix completion Most common assumpfon: Low rank decomposifon (rank r) y ij u i v j y ij = u i v j ( ) dot prodcut in r dim space

6 r dimensional space Geometric Intuition (r: the rank of the decomposifon) u a U Movies he doesn t like V T Movies he likes

7 r dimensional space Geometric Intuition (r: the rank of the decomposifon) u b u a U V T

8 Tensor data completion Tensor = MulF dimensional array Beyond 2D Movie preference + Fme / context / acfon Users Movies

9 Tensor data completion Tensor = MulF dimensional array Beyond 2D Climate monitoring temperature humidity rainfall

10 Tensor data completion Tensor = MulF dimensional array Beyond 2D Neuroscience (brain imaging) Sensors Time

11 Rest of this talk CompuFng low rank matrix decomposifon Generalizing from matrix to tensor Analyzing the performance StaFsFcal learning theory

12 CompuFng low rank matrix decomposifon

13 Computing low-rank decomposition If all entries are observed (no missing entries) m Given Y, compute singular value decomposifon (SVD) n r Y m Ur Σ r V T r where U, V: Orthogonal (U T U=I, V T V=I) Σ = ( σ1... σ r ) n σ j : jth largest singular value

14 r m U Tolerating missings OpFmizaFon problem minimize U,V r n V (ij) Ω (y ij u i v j ) 2 Set of observed index pairs

15 Tolerating missings OpFmizaFon problem r m U minimize U,V r n V v Non convex! (ij) Ω (y ij u i v j ) 2 (1,1) u ( 1, 1)

16 Tolerating missings OpFmizaFon problem SFll non convex! minimize W subject to (ij) Ω (y ij w ij ) 2, rank(w ) r Rank constraint is NP hard

17 Convex relaxation of rank Scha-en p norm (to the pth power) W p S p := r j=1 σp j (W ) σ j (W ) : jth largest singular value W p S p p 0 rank(w ) %! " # $ &!!!"!# $ # "! & '(' $)$# '(' p=1 is the Fghtest $)* '(' ( " convex relaxafon (also known as trace norm / nuclear norm)

18 Tolerating missings OpFmizaFon problem Convex relaxafon minimize W subject to Scha-en 1 norm (nuclear norm, trace norm) (ij) Ω W S1 τ Cf. Lasso (L 1 norm) for variable selecfon = linear sum of abs. coefficients (y ij w ij ) 2, W S1 = r j=1 σ j(w ) σ j (W ) : jth largest singular value

19 Take home messages Rank constrained minimizafon is hard to solve (non convex and NP hard) Can be relaxed into a tractable convex problem using Scha-en 1 norm.

20 How about tensors? How to define tensor rank? How related to matrix rank?

21 Ranf of a tensor DefiniFon. Let X R n 1 n K (Kth order tensor) The smallest number R such that the given tensor X is wri-en as X = R r=1 A r where A r = is rank one. can be wri-en as an outer product of K vectors Called CP (CANDECOMPO/PARAFAC) decomposifon Bad news: NP hard to compute the rank R even for a fully observed X.

22 Bad news 2: Tensor rank is not closed X is rank 3 X = a 1 b 1 c 2 + a 1 b 2 c 1 + a 2 b 1 c 1 Y is rank 2 Kolda & Bader 2009 Y = α(a α a 2) (b α b 2) (c α c 2) αa 1 b 1 c 1 X Y F 0 (α )

23 Tucker decomposition [Tucker 66] ( X ijk = r 1 r 2 a=1 b=1 r 3 c=1 = C abc U (1) ia U (2) jb U (3) kc ) Also known as higher order SVD [De Lathauwer+00] Rank (r 1,r 2,r 3 ) can be computed in polynomial Fme using unfolding operafons.

24 Mode-k unfoldings (matricization)

25 Computing Tucker rank For each k=1,,k Compute the mode k unfolding X (k) Compute the (matrix) rank of X (k) r k = rank(x (k) ) Unfolding Tensor X is low rank in the kth mode Folding Matrix X (k) is low rank (as a matrix)

26 Computing Tucker rank For each k=1,,k Compute the mode k unfolding X (k) Compute the (matrix) rank of X (k) r k = rank(x (k) ) Difference between Tensor rank and Tucker rank Tensor rank is a single number R (may not be easy to compute) Tucker rank is defined for each mode (easy to compute) CP decomp is a special case of Tucker decomp with R=r 1 =r 2 = =r K and diagonal core C = R R R

27 Basic idea We know how to do matrix complefon with Scha-en 1 norm (tractable convex opfmizafon) We know how to compute Tucker rank (=the rank of the mode k unfolding) Scha-en 1 norm + Unfolding = Convex, tractable tensor complefon

28 Overlapped Schatten 1-norm for Tensors W := 1 K W S1 (k) K S1 k=1 Measures the overall low rank ness (not just a single mode)

29 Convex Tensor Estimation Matrix EsFmaFon of lowrank matrix (hard) Tensor EsFmaFon of lowrank tensor (hard) Convex relaxafon Convex relaxafon Scha-en 1 norm minimizafon (tractable) [Fazel, Hindi, Boyd 01] Generalize Overlapped Scha-en 1 norm minimizafon [Liu+09, Signoretto+10, Tomioka+10, Gandy+11]

30 #!! #!!% Empirical performance Tensor complefon result [Tomioka et al. 2010] 4 71B8C'!D'!D$!E4-.3FC)D*D+ G2398D =H4I323/2398DJ KL01<1B ;8-.3/8!!"#!"$!"%!"&!"'!"(!")!"*!"+ #,-./ :48;8<8307 (No noise) M/(n 1 n 2 n 3 ) 4

31 Analyzing the performance of convex tensor decomposifon

32 ObservaFon model Problem setting y i = X i, W + ɛ i (i =1,..., M) W true tensor rank (r 1,...,r K ) ɛ i Gaussian noise Example (tensor complefon) n 3 n 2 X 1 = n1

33 ObservaFon model Problem setting y i = X i, W + ɛ i (i =1,..., M) W true tensor rank (r 1,...,r K ) ɛ i Gaussian noise Example (tensor complefon) n 3 n 2 X 2 = n1

34 ObservaFon model Problem setting y i = X i, W + ɛ i (i =1,..., M) W true tensor rank (r 1,...,r K ) ɛ i Gaussian noise Example (tensor complefon) n 3 n 2 X 3 = n1

35 ObservaFon model Problem setting y i = X i, W + ɛ i (i =1,..., M) W true tensor rank (r 1,...,r K ) ɛ i Gaussian noise Example (tensor complefon) n 3 n 2 X 4 = n1 and so on

36 ObservaFon model OpFmizaFon Ŵ = Problem setting y i = X i, W + ɛ i (i =1,..., M) argmin W R n 1 n K W ɛ i true tensor rank (r 1,...,r K ) Gaussian noise N(0,σ 2 ) ( 1 2M y X(W) λ M W S1 ) (N = K k=1 n k) X : R N R M X(W) = ( X 1, W,..., X M, W )

37 Analysis objective We would like to show something like Mean squared error EsFmated tensor Ŵ W 2 N True lowrank tensor F The size The rank Number of samples M ( ) c(n, r) O p M n =(n 1,..., n K ) r =(r 1,..., r K )

38 Theorem: random Gauss design Assume elements of X i are drown iid from standard normal distribufon Moreover #samples (M) #variables (N) c 1 n 1 1/2 r 1/2 r n n 1 1/2 := ( 1 K K k=1 1/nk ) 2, r 1/2 := ( 1 K K k=1 rk ) 2

39 Theorem: random Gauss design Assume elements of X i are drown iid from standard normal distribufon Moreover #samples (M) #variables (N) c 1 n 1 1/2 r 1/2 r n Convergence! Ŵ W 2 F N n 1 1/2 := ( ( σ 2 n 1 ) 1/2 r 1/2 O p M 1 K K k=1 1/nk ) 2, r 1/2 := ( 1 K K k=1 rk ) 2

40 What we want to derive: Ŵ W 2 N Proof idea Since Ŵ minimizes the objecfve, EsFmated tensor Obj(Ŵ) Obj(W ) It is not so hard to see: X (ɛ) = F True lowrank tensor 1 2M X(Ŵ W ) 2 2 X (ɛ)/m, Ŵ W + λm Ŵ W S1 ( ) c(n, r) O p M M ɛ i X i i=1

41 EsFmated tensor Proof outline (1/3) True lowrank tensor 1 2M X(Ŵ W ) 2 2 X (ɛ)/m, Ŵ W + λm Ŵ W S1 Inequality 1: upper bound the dot product X (ɛ)/m, Ŵ W O p ( σ 2 N n 1 1/2 M Ŵ W S1 ) (opfmizafon duality / random matrix theory)

42 Proof outline (1/3) EsFmated tensor True lowrank tensor 1 2M X(Ŵ W ) 2 2 ( σ 2 N n 1 1/2 M + λ M ) Ŵ W S 1 Inequality 1: upper bound the dot product X (ɛ)/m, Ŵ W O p ( σ 2 N n 1 1/2 M Ŵ W S1 ) Trade off between OpFmal reg. const and λ M O p ( σ 2 N n 1 1/2 M )

43 Proof outline (2/3) EsFmated True lowrank tensor tensor 1 2M X(Ŵ W ) 2 2 σ 2 N n 1 1/2 M Inequality 2: relate the scha-en 1 norm with the Frobenius norm Ŵ W S1 Ŵ W S1 r 1/2 Ŵ W F (relafon between L1 and L2 norm)

44 EsFmated tensor Proof outline (2/3) True lowrank tensor 1 2M X(Ŵ W ) 2 2 Inequality 2: relate the scha-en 1 norm with the Frobenius norm σ 2 N n 1 1/2 r 1/2 M Ŵ W S1 r 1/2 Ŵ W F (relafon between L1 and L2 norm) Ŵ W F

45 EsFmated tensor Proof outline (3/3) Inequality 3: lower bound the les hand side κ True lowrank tensor 1 2M X(Ŵ W ) 2 2 Ŵ W 2 σ 2 N n 1 1/2 r 1/2 M F 1 M X(Ŵ W ) 2 2 Ŵ W F If #samples (M) #variables (N) c 1 n 1 1/2 r 1/2 (Gordon Slepian Theorem in Gaussian process theory)

46 EsFmated tensor κ Proof outline (3/3) Ŵ W 2 Inequality 3: lower bound the les hand side κ True lowrank tensor F Ŵ W 2 σ 2 N n 1 1/2 r 1/2 M F 1 M X(Ŵ W ) 2 2 Ŵ W F If #samples (M) #variables (N) c 1 n 1 1/2 r 1/2 (Gordon Slepian Theorem in Gaussian process theory)

47 Back to the theorem statement Assume elements of X i are drown iid from standard normal distribufon Moreover Convergence! Ŵ W 2 F N NoFce: #samples (M) #variables (N) c 1 n 1 1/2 r 1/2 r n n 1 1/2 := ( σ 2 n 1 ) 1/2 r 1/2 O p M ( 1 K K k=1 1/nk ) 2, r 1/2 := ( 1 K K k=1 Sample size condifon independent of noise σ 2. Bound RHS proporfonal to σ 2. Threshold behavior in the limit σ 2 0 rk ) 2

48 : /015**/* '!! '!!( 1 Tensor completion results!!"#!"$!"%!"& ' )*+,-./01/21/345* #samples (M) #variables (N) 7*,89.)32,92:**)*;<!"!' 1 '!"&!"%!"$!"# ;/065<1=>1&1?@ ;/65<1=$!1?1>@ AB-.9.C+-./01-/85*+0,5 =./0<>?!2?!2#!@ =./0<>'!!2'!!2?!@! 2!!"#!"$!"%!"& 2 ()*+,-./012*,342553!' 55 '6# 55*55 '6#

49 Including 4 th order tensors ' 2 7*,89.)32,920**:;!"!'!"&!"%!"$!"# Gap between necessity & sufficiency? <./0;=>!2>!2#!? <./0;='!!2'!!2>!? <./0;=>!2>!2#!2'!? <./0;='!!2'!!2#!2'!?! 2!!"#!"$!"%!"& ' ()*+,-./012*,342553!' 55 55*55 '6# '6#

50 9,.:;0+54.;42,,<=!"!# #!")!"(!"& Matrix / tensor completion K = 2 >012=?'!4$!@!"$ >012=?#!!4&!@ >012=?$'!4$!!@! 4!!"#!"$!"%!"&!"'!"( *+,-./01234, !# 77 #8$ 77,77 #8$ 4 7*,89.)32,920**:;!"!' '!"&!"%!"$!"# K 3 <./0;=>!2>!2#!? <./0;='!!2'!!2>!? <./0;=>!2>!2#!2'!? <./0;='!!2'!!2#!2'!?! 2!!"#!"$!"%!"& ' ()*+,-./012*,342553!' 55 55*55 '6# '6# 2

51 Conclusion Many real world problems can be cast into the form of tensor data analysis. Convex opfmizafon is a useful tool also for the analysis of higher order tensors. Proposed a convex tensor decomposifon algorithm with performance guarantee Normalized rank predicts empirical scaling behavior well

52 Issues Why matrix complefon more difficult than tensor complefon? How big the gap between necessity and sufficiency? Random Gaussian design tensor complefon Incoherence (Candes & Recht 09) Spikiness (Negahban et al 10) When only some modes are low rank Scha-en 1 norm is too strong Mixture approach E.g. Mode 1, 4 is low rank but the rest is not (combinatorial problem)

53 T h a n k y o u!

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis Lecture 7: Matrix completion Yuejie Chi The Ohio State University Page 1 Reference Guaranteed Minimum-Rank Solutions of Linear

More information

Some tensor decomposition methods for machine learning

Some tensor decomposition methods for machine learning Some tensor decomposition methods for machine learning Massimiliano Pontil Istituto Italiano di Tecnologia and University College London 16 August 2016 1 / 36 Outline Problem and motivation Tucker decomposition

More information

Tensor Decomposition with Smoothness (ICML2017)

Tensor Decomposition with Smoothness (ICML2017) Tensor Decomposition with Smoothness (ICML2017) Masaaki Imaizumi 1 Kohei Hayashi 2,3 1 Institute of Statistical Mathematics 2 National Institute of Advanced Industrial Science and Technology 3 RIKEN Center

More information

BEYOND MATRIX COMPLETION

BEYOND MATRIX COMPLETION BEYOND MATRIX COMPLETION ANKUR MOITRA MASSACHUSETTS INSTITUTE OF TECHNOLOGY Based on joint work with Boaz Barak (MSR) Part I: RecommendaIon systems and parially observed matrices THE NETFLIX PROBLEM movies

More information

Spectral k-support Norm Regularization

Spectral k-support Norm Regularization Spectral k-support Norm Regularization Andrew McDonald Department of Computer Science, UCL (Joint work with Massimiliano Pontil and Dimitris Stamos) 25 March, 2015 1 / 19 Problem: Matrix Completion Goal:

More information

https://goo.gl/kfxweg KYOTO UNIVERSITY Statistical Machine Learning Theory Sparsity Hisashi Kashima kashima@i.kyoto-u.ac.jp DEPARTMENT OF INTELLIGENCE SCIENCE AND TECHNOLOGY 1 KYOTO UNIVERSITY Topics:

More information

Collaborative Filtering Matrix Completion Alternating Least Squares

Collaborative Filtering Matrix Completion Alternating Least Squares Case Study 4: Collaborative Filtering Collaborative Filtering Matrix Completion Alternating Least Squares Machine Learning for Big Data CSE547/STAT548, University of Washington Sham Kakade May 19, 2016

More information

Adaptive one-bit matrix completion

Adaptive one-bit matrix completion Adaptive one-bit matrix completion Joseph Salmon Télécom Paristech, Institut Mines-Télécom Joint work with Jean Lafond (Télécom Paristech) Olga Klopp (Crest / MODAL X, Université Paris Ouest) Éric Moulines

More information

Sparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28

Sparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28 Sparsity Models Tong Zhang Rutgers University T. Zhang (Rutgers) Sparsity Models 1 / 28 Topics Standard sparse regression model algorithms: convex relaxation and greedy algorithm sparse recovery analysis:

More information

Computational Linear Algebra

Computational Linear Algebra Computational Linear Algebra PD Dr. rer. nat. habil. Ralf-Peter Mundani Computation in Engineering / BGU Scientific Computing in Computer Science / INF Winter Term 2018/19 Part 6: Some Other Stuff PD Dr.

More information

Rank Determination for Low-Rank Data Completion

Rank Determination for Low-Rank Data Completion Journal of Machine Learning Research 18 017) 1-9 Submitted 7/17; Revised 8/17; Published 9/17 Rank Determination for Low-Rank Data Completion Morteza Ashraphijuo Columbia University New York, NY 1007,

More information

CSC 576: Variants of Sparse Learning

CSC 576: Variants of Sparse Learning CSC 576: Variants of Sparse Learning Ji Liu Department of Computer Science, University of Rochester October 27, 205 Introduction Our previous note basically suggests using l norm to enforce sparsity in

More information

Collaborative Filtering

Collaborative Filtering Case Study 4: Collaborative Filtering Collaborative Filtering Matrix Completion Alternating Least Squares Machine Learning/Statistics for Big Data CSE599C1/STAT592, University of Washington Carlos Guestrin

More information

BEYOND MATRIX COMPLETION

BEYOND MATRIX COMPLETION BEYOND MATRIX COMPLETION ANKUR MOITRA MASSACHUSETTS INSTITUTE OF TECHNOLOGY Based on joint work with Boaz Barak (Harvard) Part I: RecommendaJon systems and parjally observed matrices THE NETFLIX PROBLEM

More information

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and Inference Low-rank matrix recovery via convex relaxations Yuejie Chi Department of Electrical and Computer Engineering Spring

More information

Matrix Factorizations: A Tale of Two Norms

Matrix Factorizations: A Tale of Two Norms Matrix Factorizations: A Tale of Two Norms Nati Srebro Toyota Technological Institute Chicago Maximum Margin Matrix Factorization S, Jason Rennie, Tommi Jaakkola (MIT), NIPS 2004 Rank, Trace-Norm and Max-Norm

More information

A Fast Augmented Lagrangian Algorithm for Learning Low-Rank Matrices

A Fast Augmented Lagrangian Algorithm for Learning Low-Rank Matrices A Fast Augmented Lagrangian Algorithm for Learning Low-Rank Matrices Ryota Tomioka 1, Taiji Suzuki 1, Masashi Sugiyama 2, Hisashi Kashima 1 1 The University of Tokyo 2 Tokyo Institute of Technology 2010-06-22

More information

Provable Alternating Minimization Methods for Non-convex Optimization

Provable Alternating Minimization Methods for Non-convex Optimization Provable Alternating Minimization Methods for Non-convex Optimization Prateek Jain Microsoft Research, India Joint work with Praneeth Netrapalli, Sujay Sanghavi, Alekh Agarwal, Animashree Anandkumar, Rashish

More information

Linear dimensionality reduction for data analysis

Linear dimensionality reduction for data analysis Linear dimensionality reduction for data analysis Nicolas Gillis Joint work with Robert Luce, François Glineur, Stephen Vavasis, Robert Plemmons, Gabriella Casalino The setup Dimensionality reduction for

More information

Linear Algebra (Review) Volker Tresp 2017

Linear Algebra (Review) Volker Tresp 2017 Linear Algebra (Review) Volker Tresp 2017 1 Vectors k is a scalar (a number) c is a column vector. Thus in two dimensions, c = ( c1 c 2 ) (Advanced: More precisely, a vector is defined in a vector space.

More information

A Characterization of Sampling Patterns for Union of Low-Rank Subspaces Retrieval Problem

A Characterization of Sampling Patterns for Union of Low-Rank Subspaces Retrieval Problem A Characterization of Sampling Patterns for Union of Low-Rank Subspaces Retrieval Problem Morteza Ashraphijuo Columbia University ashraphijuo@ee.columbia.edu Xiaodong Wang Columbia University wangx@ee.columbia.edu

More information

Matrix Completion: Fundamental Limits and Efficient Algorithms

Matrix Completion: Fundamental Limits and Efficient Algorithms Matrix Completion: Fundamental Limits and Efficient Algorithms Sewoong Oh PhD Defense Stanford University July 23, 2010 1 / 33 Matrix completion Find the missing entries in a huge data matrix 2 / 33 Example

More information

1 Regression with High Dimensional Data

1 Regression with High Dimensional Data 6.883 Learning with Combinatorial Structure ote for Lecture 11 Instructor: Prof. Stefanie Jegelka Scribe: Xuhong Zhang 1 Regression with High Dimensional Data Consider the following regression problem:

More information

Linear Algebra (Review) Volker Tresp 2018

Linear Algebra (Review) Volker Tresp 2018 Linear Algebra (Review) Volker Tresp 2018 1 Vectors k, M, N are scalars A one-dimensional array c is a column vector. Thus in two dimensions, ( ) c1 c = c 2 c i is the i-th component of c c T = (c 1, c

More information

Matrix Completion from Fewer Entries

Matrix Completion from Fewer Entries from Fewer Entries Stanford University March 30, 2009 Outline The problem, a look at the data, and some results (slides) 2 Proofs (blackboard) arxiv:090.350 The problem, a look at the data, and some results

More information

Structured matrix factorizations. Example: Eigenfaces

Structured matrix factorizations. Example: Eigenfaces Structured matrix factorizations Example: Eigenfaces An extremely large variety of interesting and important problems in machine learning can be formulated as: Given a matrix, find a matrix and a matrix

More information

Recovering any low-rank matrix, provably

Recovering any low-rank matrix, provably Recovering any low-rank matrix, provably Rachel Ward University of Texas at Austin October, 2014 Joint work with Yudong Chen (U.C. Berkeley), Srinadh Bhojanapalli and Sujay Sanghavi (U.T. Austin) Matrix

More information

Tensor Low-Rank Completion and Invariance of the Tucker Core

Tensor Low-Rank Completion and Invariance of the Tucker Core Tensor Low-Rank Completion and Invariance of the Tucker Core Shuzhong Zhang Department of Industrial & Systems Engineering University of Minnesota zhangs@umn.edu Joint work with Bo JIANG, Shiqian MA, and

More information

Lecture Notes 10: Matrix Factorization

Lecture Notes 10: Matrix Factorization Optimization-based data analysis Fall 207 Lecture Notes 0: Matrix Factorization Low-rank models. Rank- model Consider the problem of modeling a quantity y[i, j] that depends on two indices i and j. To

More information

Maximum Margin Matrix Factorization

Maximum Margin Matrix Factorization Maximum Margin Matrix Factorization Nati Srebro Toyota Technological Institute Chicago Joint work with Noga Alon Tel-Aviv Yonatan Amit Hebrew U Alex d Aspremont Princeton Michael Fink Hebrew U Tommi Jaakkola

More information

GI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis. Massimiliano Pontil

GI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis. Massimiliano Pontil GI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis Massimiliano Pontil 1 Today s plan SVD and principal component analysis (PCA) Connection

More information

Generalized Higher-Order Tensor Decomposition via Parallel ADMM

Generalized Higher-Order Tensor Decomposition via Parallel ADMM Proceedings of the Twenty-Eighth AAAI Conference on Artificial Intelligence Generalized Higher-Order Tensor Decomposition via Parallel ADMM Fanhua Shang 1, Yuanyuan Liu 2, James Cheng 1 1 Department of

More information

Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms

Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms François Caron Department of Statistics, Oxford STATLEARN 2014, Paris April 7, 2014 Joint work with Adrien Todeschini,

More information

Low-rank Matrix Completion with Noisy Observations: a Quantitative Comparison

Low-rank Matrix Completion with Noisy Observations: a Quantitative Comparison Low-rank Matrix Completion with Noisy Observations: a Quantitative Comparison Raghunandan H. Keshavan, Andrea Montanari and Sewoong Oh Electrical Engineering and Statistics Department Stanford University,

More information

Matrix completion: Fundamental limits and efficient algorithms. Sewoong Oh Stanford University

Matrix completion: Fundamental limits and efficient algorithms. Sewoong Oh Stanford University Matrix completion: Fundamental limits and efficient algorithms Sewoong Oh Stanford University 1 / 35 Low-rank matrix completion Low-rank Data Matrix Sparse Sampled Matrix Complete the matrix from small

More information

Robust Tensor Completion and its Applications. Michael K. Ng Department of Mathematics Hong Kong Baptist University

Robust Tensor Completion and its Applications. Michael K. Ng Department of Mathematics Hong Kong Baptist University Robust Tensor Completion and its Applications Michael K. Ng Department of Mathematics Hong Kong Baptist University mng@math.hkbu.edu.hk Example Low-dimensional Structure Data in many real applications

More information

Matrix Rank Minimization with Applications

Matrix Rank Minimization with Applications Matrix Rank Minimization with Applications Maryam Fazel Haitham Hindi Stephen Boyd Information Systems Lab Electrical Engineering Department Stanford University 8/2001 ACC 01 Outline Rank Minimization

More information

Robust Principal Component Analysis

Robust Principal Component Analysis ELE 538B: Mathematics of High-Dimensional Data Robust Principal Component Analysis Yuxin Chen Princeton University, Fall 2018 Disentangling sparse and low-rank matrices Suppose we are given a matrix M

More information

Conditions for Robust Principal Component Analysis

Conditions for Robust Principal Component Analysis Rose-Hulman Undergraduate Mathematics Journal Volume 12 Issue 2 Article 9 Conditions for Robust Principal Component Analysis Michael Hornstein Stanford University, mdhornstein@gmail.com Follow this and

More information

Low Rank Matrix Completion Formulation and Algorithm

Low Rank Matrix Completion Formulation and Algorithm 1 2 Low Rank Matrix Completion and Algorithm Jian Zhang Department of Computer Science, ETH Zurich zhangjianthu@gmail.com March 25, 2014 Movie Rating 1 2 Critic A 5 5 Critic B 6 5 Jian 9 8 Kind Guy B 9

More information

Using SVD to Recommend Movies

Using SVD to Recommend Movies Michael Percy University of California, Santa Cruz Last update: December 12, 2009 Last update: December 12, 2009 1 / Outline 1 Introduction 2 Singular Value Decomposition 3 Experiments 4 Conclusion Last

More information

15-780: LinearProgramming

15-780: LinearProgramming 15-780: LinearProgramming J. Zico Kolter February 1-3, 2016 1 Outline Introduction Some linear algebra review Linear programming Simplex algorithm Duality and dual simplex 2 Outline Introduction Some linear

More information

Rank minimization via the γ 2 norm

Rank minimization via the γ 2 norm Rank minimization via the γ 2 norm Troy Lee Columbia University Adi Shraibman Weizmann Institute Rank Minimization Problem Consider the following problem min X rank(x) A i, X b i for i = 1,..., k Arises

More information

The convex algebraic geometry of rank minimization

The convex algebraic geometry of rank minimization The convex algebraic geometry of rank minimization Pablo A. Parrilo Laboratory for Information and Decision Systems Massachusetts Institute of Technology International Symposium on Mathematical Programming

More information

sparse and low-rank tensor recovery Cubic-Sketching

sparse and low-rank tensor recovery Cubic-Sketching Sparse and Low-Ran Tensor Recovery via Cubic-Setching Guang Cheng Department of Statistics Purdue University www.science.purdue.edu/bigdata CCAM@Purdue Math Oct. 27, 2017 Joint wor with Botao Hao and Anru

More information

Low-Rank Matrix Recovery

Low-Rank Matrix Recovery ELE 538B: Mathematics of High-Dimensional Data Low-Rank Matrix Recovery Yuxin Chen Princeton University, Fall 2018 Outline Motivation Problem setup Nuclear norm minimization RIP and low-rank matrix recovery

More information

An iterative hard thresholding estimator for low rank matrix recovery

An iterative hard thresholding estimator for low rank matrix recovery An iterative hard thresholding estimator for low rank matrix recovery Alexandra Carpentier - based on a joint work with Arlene K.Y. Kim Statistical Laboratory, Department of Pure Mathematics and Mathematical

More information

1-Bit Matrix Completion

1-Bit Matrix Completion 1-Bit Matrix Completion Mark A. Davenport School of Electrical and Computer Engineering Georgia Institute of Technology Yaniv Plan Mary Wootters Ewout van den Berg Matrix Completion d When is it possible

More information

Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs

Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs Raphael Louca & Eilyan Bitar School of Electrical and Computer Engineering American Control Conference (ACC) Chicago,

More information

Lecture 3. Random Fourier measurements

Lecture 3. Random Fourier measurements Lecture 3. Random Fourier measurements 1 Sampling from Fourier matrices 2 Law of Large Numbers and its operator-valued versions 3 Frames. Rudelson s Selection Theorem Sampling from Fourier matrices Our

More information

Binary matrix completion

Binary matrix completion Binary matrix completion Yaniv Plan University of Michigan SAMSI, LDHD workshop, 2013 Joint work with (a) Mark Davenport (b) Ewout van den Berg (c) Mary Wootters Yaniv Plan (U. Mich.) Binary matrix completion

More information

Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms

Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms Adrien Todeschini Inria Bordeaux JdS 2014, Rennes Aug. 2014 Joint work with François Caron (Univ. Oxford), Marie

More information

An Extended Frank-Wolfe Method, with Application to Low-Rank Matrix Completion

An Extended Frank-Wolfe Method, with Application to Low-Rank Matrix Completion An Extended Frank-Wolfe Method, with Application to Low-Rank Matrix Completion Robert M. Freund, MIT joint with Paul Grigas (UC Berkeley) and Rahul Mazumder (MIT) CDC, December 2016 1 Outline of Topics

More information

Rank, Trace-Norm & Max-Norm

Rank, Trace-Norm & Max-Norm Rank, Trace-Norm & Max-Norm as measures of matrix complexity Nati Srebro University of Toronto Adi Shraibman Hebrew University Matrix Learning users movies 2 1 4 5 5 4? 1 3 3 5 2 4? 5 3? 4 1 3 5 2 1? 4

More information

Lecture 20: November 1st

Lecture 20: November 1st 10-725: Optimization Fall 2012 Lecture 20: November 1st Lecturer: Geoff Gordon Scribes: Xiaolong Shen, Alex Beutel Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These notes have not

More information

Tensor Analysis. Topics in Data Mining Fall Bruno Ribeiro

Tensor Analysis. Topics in Data Mining Fall Bruno Ribeiro Tensor Analysis Topics in Data Mining Fall 2015 Bruno Ribeiro Tensor Basics But First 2 Mining Matrices 3 Singular Value Decomposition (SVD) } X(i,j) = value of user i for property j i 2 j 5 X(Alice, cholesterol)

More information

1-Bit Matrix Completion

1-Bit Matrix Completion 1-Bit Matrix Completion Mark A. Davenport School of Electrical and Computer Engineering Georgia Institute of Technology Yaniv Plan Mary Wootters Ewout van den Berg Matrix Completion d When is it possible

More information

Factor Matrix Trace Norm Minimization for Low-Rank Tensor Completion

Factor Matrix Trace Norm Minimization for Low-Rank Tensor Completion Factor Matrix Trace Norm Minimization for Low-Rank Tensor Completion Yuanyuan Liu Fanhua Shang Hong Cheng James Cheng Hanghang Tong Abstract Most existing low-n-rank imization algorithms for tensor completion

More information

Tensor Robust Principal Component Analysis: Exact Recovery of Corrupted Low-Rank Tensors via Convex Optimization

Tensor Robust Principal Component Analysis: Exact Recovery of Corrupted Low-Rank Tensors via Convex Optimization Tensor Robust Principal Component Analysis: Exact Recovery of Corrupted Low-Rank Tensors via Convex Optimization Canyi Lu 1, Jiashi Feng 1, Yudong Chen, Wei Liu 3, Zhouchen Lin 4,5,, Shuicheng Yan 6,1

More information

Matrix Completion for Structured Observations

Matrix Completion for Structured Observations Matrix Completion for Structured Observations Denali Molitor Department of Mathematics University of California, Los ngeles Los ngeles, C 90095, US Email: dmolitor@math.ucla.edu Deanna Needell Department

More information

CS264: Beyond Worst-Case Analysis Lecture #15: Topic Modeling and Nonnegative Matrix Factorization

CS264: Beyond Worst-Case Analysis Lecture #15: Topic Modeling and Nonnegative Matrix Factorization CS264: Beyond Worst-Case Analysis Lecture #15: Topic Modeling and Nonnegative Matrix Factorization Tim Roughgarden February 28, 2017 1 Preamble This lecture fulfills a promise made back in Lecture #1,

More information

Basic Concepts in Linear Algebra

Basic Concepts in Linear Algebra Basic Concepts in Linear Algebra Grady B Wright Department of Mathematics Boise State University February 2, 2015 Grady B Wright Linear Algebra Basics February 2, 2015 1 / 39 Numerical Linear Algebra Linear

More information

Information-Theoretic Limits of Matrix Completion

Information-Theoretic Limits of Matrix Completion Information-Theoretic Limits of Matrix Completion Erwin Riegler, David Stotz, and Helmut Bölcskei Dept. IT & EE, ETH Zurich, Switzerland Email: {eriegler, dstotz, boelcskei}@nari.ee.ethz.ch Abstract We

More information

Compressed Sensing and Robust Recovery of Low Rank Matrices

Compressed Sensing and Robust Recovery of Low Rank Matrices Compressed Sensing and Robust Recovery of Low Rank Matrices M. Fazel, E. Candès, B. Recht, P. Parrilo Electrical Engineering, University of Washington Applied and Computational Mathematics Dept., Caltech

More information

CHAPTER 11. A Revision. 1. The Computers and Numbers therein

CHAPTER 11. A Revision. 1. The Computers and Numbers therein CHAPTER A Revision. The Computers and Numbers therein Traditional computer science begins with a finite alphabet. By stringing elements of the alphabet one after another, one obtains strings. A set of

More information

Lecture 8: Linear Algebra Background

Lecture 8: Linear Algebra Background CSE 521: Design and Analysis of Algorithms I Winter 2017 Lecture 8: Linear Algebra Background Lecturer: Shayan Oveis Gharan 2/1/2017 Scribe: Swati Padmanabhan Disclaimer: These notes have not been subjected

More information

Approximate Low-Rank Tensor Learning

Approximate Low-Rank Tensor Learning Approximate Low-Rank Tensor Learning Yaoliang Yu Dept of Machine Learning Carnegie Mellon University yaoliang@cs.cmu.edu Hao Cheng Dept of Electric Engineering University of Washington kelvinwsch@gmail.com

More information

Review of Basic Concepts in Linear Algebra

Review of Basic Concepts in Linear Algebra Review of Basic Concepts in Linear Algebra Grady B Wright Department of Mathematics Boise State University September 7, 2017 Math 565 Linear Algebra Review September 7, 2017 1 / 40 Numerical Linear Algebra

More information

High Dimensional Covariance and Precision Matrix Estimation

High Dimensional Covariance and Precision Matrix Estimation High Dimensional Covariance and Precision Matrix Estimation Wei Wang Washington University in St. Louis Thursday 23 rd February, 2017 Wei Wang (Washington University in St. Louis) High Dimensional Covariance

More information

Information Retrieval

Information Retrieval Introduction to Information CS276: Information and Web Search Christopher Manning and Pandu Nayak Lecture 13: Latent Semantic Indexing Ch. 18 Today s topic Latent Semantic Indexing Term-document matrices

More information

CS598 Machine Learning in Computational Biology (Lecture 5: Matrix - part 2) Professor Jian Peng Teaching Assistant: Rongda Zhu

CS598 Machine Learning in Computational Biology (Lecture 5: Matrix - part 2) Professor Jian Peng Teaching Assistant: Rongda Zhu CS598 Machine Learning in Computational Biology (Lecture 5: Matrix - part 2) Professor Jian Peng Teaching Assistant: Rongda Zhu Feature engineering is hard 1. Extract informative features from domain knowledge

More information

arxiv: v1 [math.oc] 11 Jun 2009

arxiv: v1 [math.oc] 11 Jun 2009 RANK-SPARSITY INCOHERENCE FOR MATRIX DECOMPOSITION VENKAT CHANDRASEKARAN, SUJAY SANGHAVI, PABLO A. PARRILO, S. WILLSKY AND ALAN arxiv:0906.2220v1 [math.oc] 11 Jun 2009 Abstract. Suppose we are given a

More information

Sparse and Low-Rank Matrix Decompositions

Sparse and Low-Rank Matrix Decompositions Forty-Seventh Annual Allerton Conference Allerton House, UIUC, Illinois, USA September 30 - October 2, 2009 Sparse and Low-Rank Matrix Decompositions Venkat Chandrasekaran, Sujay Sanghavi, Pablo A. Parrilo,

More information

An Introduction to Sparse Approximation

An Introduction to Sparse Approximation An Introduction to Sparse Approximation Anna C. Gilbert Department of Mathematics University of Michigan Basic image/signal/data compression: transform coding Approximate signals sparsely Compress images,

More information

The Stability of Low-Rank Matrix Reconstruction: a Constrained Singular Value Perspective

The Stability of Low-Rank Matrix Reconstruction: a Constrained Singular Value Perspective Forty-Eighth Annual Allerton Conference Allerton House UIUC Illinois USA September 9 - October 1 010 The Stability of Low-Rank Matrix Reconstruction: a Constrained Singular Value Perspective Gongguo Tang

More information

1-Bit Matrix Completion

1-Bit Matrix Completion 1-Bit Matrix Completion Mark A. Davenport School of Electrical and Computer Engineering Georgia Institute of Technology Yaniv Plan Mary Wootters Ewout van den Berg Matrix Completion d When is it possible

More information

Introduction to Tensors. 8 May 2014

Introduction to Tensors. 8 May 2014 Introduction to Tensors 8 May 2014 Introduction to Tensors What is a tensor? Basic Operations CP Decompositions and Tensor Rank Matricization and Computing the CP Dear Tullio,! I admire the elegance of

More information

Concentration-Based Guarantees for Low-Rank Matrix Reconstruction

Concentration-Based Guarantees for Low-Rank Matrix Reconstruction JMLR: Workshop and Conference Proceedings 9 20 35 339 24th Annual Conference on Learning Theory Concentration-Based Guarantees for Low-Rank Matrix Reconstruction Rina Foygel Department of Statistics, University

More information

Non-Convex Projected Gradient Descent for Generalized Low-Rank Tensor Regression

Non-Convex Projected Gradient Descent for Generalized Low-Rank Tensor Regression Non-Convex Projected Gradient Descent for Generalized Low-Rank Tensor Regression Han Chen, Garvesh Raskutti and Ming Yuan University of Wisconsin-Madison Morgridge Institute for Research and Department

More information

Large Scale Data Analysis Using Deep Learning

Large Scale Data Analysis Using Deep Learning Large Scale Data Analysis Using Deep Learning Linear Algebra U Kang Seoul National University U Kang 1 In This Lecture Overview of linear algebra (but, not a comprehensive survey) Focused on the subset

More information

Andriy Mnih and Ruslan Salakhutdinov

Andriy Mnih and Ruslan Salakhutdinov MATRIX FACTORIZATION METHODS FOR COLLABORATIVE FILTERING Andriy Mnih and Ruslan Salakhutdinov University of Toronto, Machine Learning Group 1 What is collaborative filtering? The goal of collaborative

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Instructor: Moritz Hardt Email: hardt+ee227c@berkeley.edu Graduate Instructor: Max Simchowitz Email: msimchow+ee227c@berkeley.edu

More information

Faloutsos, Tong ICDE, 2009

Faloutsos, Tong ICDE, 2009 Large Graph Mining: Patterns, Tools and Case Studies Christos Faloutsos Hanghang Tong CMU Copyright: Faloutsos, Tong (29) 2-1 Outline Part 1: Patterns Part 2: Matrix and Tensor Tools Part 3: Proximity

More information

On Tensor Train Rank Minimization: Statistical Efficiency and Scalable Algorithm

On Tensor Train Rank Minimization: Statistical Efficiency and Scalable Algorithm On Tensor Train Ran Minimization: Statistical Efficiency and Scalable Algorithm Masaai Imaizumi Institute of Statistical Mathematics RIKEN Center for Advanced Intelligence Project imaizumi@ism.ac.jp Taanori

More information

Matrix and Tensor Factorization from a Machine Learning Perspective

Matrix and Tensor Factorization from a Machine Learning Perspective Matrix and Tensor Factorization from a Machine Learning Perspective Christoph Freudenthaler Information Systems and Machine Learning Lab, University of Hildesheim Research Seminar, Vienna University of

More information

Recovery of Simultaneously Structured Models using Convex Optimization

Recovery of Simultaneously Structured Models using Convex Optimization Recovery of Simultaneously Structured Models using Convex Optimization Maryam Fazel University of Washington Joint work with: Amin Jalali (UW), Samet Oymak and Babak Hassibi (Caltech) Yonina Eldar (Technion)

More information

High-dimensional Statistics

High-dimensional Statistics High-dimensional Statistics Pradeep Ravikumar UT Austin Outline 1. High Dimensional Data : Large p, small n 2. Sparsity 3. Group Sparsity 4. Low Rank 1 Curse of Dimensionality Statistical Learning: Given

More information

High-Rank Matrix Completion and Subspace Clustering with Missing Data

High-Rank Matrix Completion and Subspace Clustering with Missing Data High-Rank Matrix Completion and Subspace Clustering with Missing Data Authors: Brian Eriksson, Laura Balzano and Robert Nowak Presentation by Wang Yuxiang 1 About the authors

More information

Knowledge Discovery and Data Mining 1 (VO) ( )

Knowledge Discovery and Data Mining 1 (VO) ( ) Knowledge Discovery and Data Mining 1 (VO) (707.003) Review of Linear Algebra Denis Helic KTI, TU Graz Oct 9, 2014 Denis Helic (KTI, TU Graz) KDDM1 Oct 9, 2014 1 / 74 Big picture: KDDM Probability Theory

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course Notes for EE7C (Spring 08): Convex Optimization and Approximation Instructor: Moritz Hardt Email: hardt+ee7c@berkeley.edu Graduate Instructor: Max Simchowitz Email: msimchow+ee7c@berkeley.edu October

More information

STA141C: Big Data & High Performance Statistical Computing

STA141C: Big Data & High Performance Statistical Computing STA141C: Big Data & High Performance Statistical Computing Numerical Linear Algebra Background Cho-Jui Hsieh UC Davis May 15, 2018 Linear Algebra Background Vectors A vector has a direction and a magnitude

More information

Tensor Decompositions for Machine Learning. G. Roeder 1. UBC Machine Learning Reading Group, June University of British Columbia

Tensor Decompositions for Machine Learning. G. Roeder 1. UBC Machine Learning Reading Group, June University of British Columbia Network Feature s Decompositions for Machine Learning 1 1 Department of Computer Science University of British Columbia UBC Machine Learning Group, June 15 2016 1/30 Contact information Network Feature

More information

SPARSE signal representations have gained popularity in recent

SPARSE signal representations have gained popularity in recent 6958 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 10, OCTOBER 2011 Blind Compressed Sensing Sivan Gleichman and Yonina C. Eldar, Senior Member, IEEE Abstract The fundamental principle underlying

More information

Tensor Decompositions and Applications

Tensor Decompositions and Applications Tamara G. Kolda and Brett W. Bader Part I September 22, 2015 What is tensor? A N-th order tensor is an element of the tensor product of N vector spaces, each of which has its own coordinate system. a =

More information

Learning with matrix and tensor based models using low-rank penalties

Learning with matrix and tensor based models using low-rank penalties Learning with matrix and tensor based models using low-rank penalties Johan Suykens KU Leuven, ESAT-SCD/SISTA Kasteelpark Arenberg 10 B-3001 Leuven (Heverlee), Belgium Email: johan.suykens@esat.kuleuven.be

More information

Signal Recovery from Permuted Observations

Signal Recovery from Permuted Observations EE381V Course Project Signal Recovery from Permuted Observations 1 Problem Shanshan Wu (sw33323) May 8th, 2015 We start with the following problem: let s R n be an unknown n-dimensional real-valued signal,

More information

Geometric Modeling Summer Semester 2010 Mathematical Tools (1)

Geometric Modeling Summer Semester 2010 Mathematical Tools (1) Geometric Modeling Summer Semester 2010 Mathematical Tools (1) Recap: Linear Algebra Today... Topics: Mathematical Background Linear algebra Analysis & differential geometry Numerical techniques Geometric

More information

Non-Convex Projected Gradient Descent for Generalized Low-Rank Tensor Regression

Non-Convex Projected Gradient Descent for Generalized Low-Rank Tensor Regression Journal of Machine Learning Research 20 (2019) 1-37 Submitted 12/16; Revised 1/18; Published 2/19 Non-Convex Projected Gradient Descent for Generalized Low-Rank Tensor Regression Han Chen Garvesh Raskutti

More information

Exact Low-rank Matrix Recovery via Nonconvex M p -Minimization

Exact Low-rank Matrix Recovery via Nonconvex M p -Minimization Exact Low-rank Matrix Recovery via Nonconvex M p -Minimization Lingchen Kong and Naihua Xiu Department of Applied Mathematics, Beijing Jiaotong University, Beijing, 100044, People s Republic of China E-mail:

More information

Optimisation Combinatoire et Convexe.

Optimisation Combinatoire et Convexe. Optimisation Combinatoire et Convexe. Low complexity models, l 1 penalties. A. d Aspremont. M1 ENS. 1/36 Today Sparsity, low complexity models. l 1 -recovery results: three approaches. Extensions: matrix

More information