Statistical Performance of Convex Tensor Decomposition
|
|
- Alban Barnett
- 5 years ago
- Views:
Transcription
1 Slides available: h-p:// tokyo.ac.jp/ryotat/tensor12kyoto.pdf Statistical Performance of Convex Tensor Decomposition Ryota Tomioka Kyoto University Perspectives in Informatics 4B Collaborators: Taiji Suzuki Kohei Hayashi Hisashi Kashima
2 Netflix challenge ( ) $1,000,000 prize Goal: Improve the performance of a video recommendafon system (predict who likes which movies) Example: Likes Star Wars and E.T., Doesn t like Minority Report. Does he like Blade Runner?
3 Matrix completion view User A User B User C User D ?? +1?? +1?? +1 1? +1 +1??? Goal: fill the missing entries!
4 Matrix completion Impossible without an assumpfon. (Missing entries can be arbitrary) problem is ill posed Most common assumpfon: Low rank decomposifon Y U V T
5 Matrix completion Most common assumpfon: Low rank decomposifon (rank r) y ij u i v j y ij = u i v j ( ) dot prodcut in r dim space
6 r dimensional space Geometric Intuition (r: the rank of the decomposifon) u a U Movies he doesn t like V T Movies he likes
7 r dimensional space Geometric Intuition (r: the rank of the decomposifon) u b u a U V T
8 Tensor data completion Tensor = MulF dimensional array Beyond 2D Movie preference + Fme / context / acfon Users Movies
9 Tensor data completion Tensor = MulF dimensional array Beyond 2D Climate monitoring temperature humidity rainfall
10 Tensor data completion Tensor = MulF dimensional array Beyond 2D Neuroscience (brain imaging) Sensors Time
11 Rest of this talk CompuFng low rank matrix decomposifon Generalizing from matrix to tensor Analyzing the performance StaFsFcal learning theory
12 CompuFng low rank matrix decomposifon
13 Computing low-rank decomposition If all entries are observed (no missing entries) m Given Y, compute singular value decomposifon (SVD) n r Y m Ur Σ r V T r where U, V: Orthogonal (U T U=I, V T V=I) Σ = ( σ1... σ r ) n σ j : jth largest singular value
14 r m U Tolerating missings OpFmizaFon problem minimize U,V r n V (ij) Ω (y ij u i v j ) 2 Set of observed index pairs
15 Tolerating missings OpFmizaFon problem r m U minimize U,V r n V v Non convex! (ij) Ω (y ij u i v j ) 2 (1,1) u ( 1, 1)
16 Tolerating missings OpFmizaFon problem SFll non convex! minimize W subject to (ij) Ω (y ij w ij ) 2, rank(w ) r Rank constraint is NP hard
17 Convex relaxation of rank Scha-en p norm (to the pth power) W p S p := r j=1 σp j (W ) σ j (W ) : jth largest singular value W p S p p 0 rank(w ) %! " # $ &!!!"!# $ # "! & '(' $)$# '(' p=1 is the Fghtest $)* '(' ( " convex relaxafon (also known as trace norm / nuclear norm)
18 Tolerating missings OpFmizaFon problem Convex relaxafon minimize W subject to Scha-en 1 norm (nuclear norm, trace norm) (ij) Ω W S1 τ Cf. Lasso (L 1 norm) for variable selecfon = linear sum of abs. coefficients (y ij w ij ) 2, W S1 = r j=1 σ j(w ) σ j (W ) : jth largest singular value
19 Take home messages Rank constrained minimizafon is hard to solve (non convex and NP hard) Can be relaxed into a tractable convex problem using Scha-en 1 norm.
20 How about tensors? How to define tensor rank? How related to matrix rank?
21 Ranf of a tensor DefiniFon. Let X R n 1 n K (Kth order tensor) The smallest number R such that the given tensor X is wri-en as X = R r=1 A r where A r = is rank one. can be wri-en as an outer product of K vectors Called CP (CANDECOMPO/PARAFAC) decomposifon Bad news: NP hard to compute the rank R even for a fully observed X.
22 Bad news 2: Tensor rank is not closed X is rank 3 X = a 1 b 1 c 2 + a 1 b 2 c 1 + a 2 b 1 c 1 Y is rank 2 Kolda & Bader 2009 Y = α(a α a 2) (b α b 2) (c α c 2) αa 1 b 1 c 1 X Y F 0 (α )
23 Tucker decomposition [Tucker 66] ( X ijk = r 1 r 2 a=1 b=1 r 3 c=1 = C abc U (1) ia U (2) jb U (3) kc ) Also known as higher order SVD [De Lathauwer+00] Rank (r 1,r 2,r 3 ) can be computed in polynomial Fme using unfolding operafons.
24 Mode-k unfoldings (matricization)
25 Computing Tucker rank For each k=1,,k Compute the mode k unfolding X (k) Compute the (matrix) rank of X (k) r k = rank(x (k) ) Unfolding Tensor X is low rank in the kth mode Folding Matrix X (k) is low rank (as a matrix)
26 Computing Tucker rank For each k=1,,k Compute the mode k unfolding X (k) Compute the (matrix) rank of X (k) r k = rank(x (k) ) Difference between Tensor rank and Tucker rank Tensor rank is a single number R (may not be easy to compute) Tucker rank is defined for each mode (easy to compute) CP decomp is a special case of Tucker decomp with R=r 1 =r 2 = =r K and diagonal core C = R R R
27 Basic idea We know how to do matrix complefon with Scha-en 1 norm (tractable convex opfmizafon) We know how to compute Tucker rank (=the rank of the mode k unfolding) Scha-en 1 norm + Unfolding = Convex, tractable tensor complefon
28 Overlapped Schatten 1-norm for Tensors W := 1 K W S1 (k) K S1 k=1 Measures the overall low rank ness (not just a single mode)
29 Convex Tensor Estimation Matrix EsFmaFon of lowrank matrix (hard) Tensor EsFmaFon of lowrank tensor (hard) Convex relaxafon Convex relaxafon Scha-en 1 norm minimizafon (tractable) [Fazel, Hindi, Boyd 01] Generalize Overlapped Scha-en 1 norm minimizafon [Liu+09, Signoretto+10, Tomioka+10, Gandy+11]
30 #!! #!!% Empirical performance Tensor complefon result [Tomioka et al. 2010] 4 71B8C'!D'!D$!E4-.3FC)D*D+ G2398D =H4I323/2398DJ KL01<1B ;8-.3/8!!"#!"$!"%!"&!"'!"(!")!"*!"+ #,-./ :48;8<8307 (No noise) M/(n 1 n 2 n 3 ) 4
31 Analyzing the performance of convex tensor decomposifon
32 ObservaFon model Problem setting y i = X i, W + ɛ i (i =1,..., M) W true tensor rank (r 1,...,r K ) ɛ i Gaussian noise Example (tensor complefon) n 3 n 2 X 1 = n1
33 ObservaFon model Problem setting y i = X i, W + ɛ i (i =1,..., M) W true tensor rank (r 1,...,r K ) ɛ i Gaussian noise Example (tensor complefon) n 3 n 2 X 2 = n1
34 ObservaFon model Problem setting y i = X i, W + ɛ i (i =1,..., M) W true tensor rank (r 1,...,r K ) ɛ i Gaussian noise Example (tensor complefon) n 3 n 2 X 3 = n1
35 ObservaFon model Problem setting y i = X i, W + ɛ i (i =1,..., M) W true tensor rank (r 1,...,r K ) ɛ i Gaussian noise Example (tensor complefon) n 3 n 2 X 4 = n1 and so on
36 ObservaFon model OpFmizaFon Ŵ = Problem setting y i = X i, W + ɛ i (i =1,..., M) argmin W R n 1 n K W ɛ i true tensor rank (r 1,...,r K ) Gaussian noise N(0,σ 2 ) ( 1 2M y X(W) λ M W S1 ) (N = K k=1 n k) X : R N R M X(W) = ( X 1, W,..., X M, W )
37 Analysis objective We would like to show something like Mean squared error EsFmated tensor Ŵ W 2 N True lowrank tensor F The size The rank Number of samples M ( ) c(n, r) O p M n =(n 1,..., n K ) r =(r 1,..., r K )
38 Theorem: random Gauss design Assume elements of X i are drown iid from standard normal distribufon Moreover #samples (M) #variables (N) c 1 n 1 1/2 r 1/2 r n n 1 1/2 := ( 1 K K k=1 1/nk ) 2, r 1/2 := ( 1 K K k=1 rk ) 2
39 Theorem: random Gauss design Assume elements of X i are drown iid from standard normal distribufon Moreover #samples (M) #variables (N) c 1 n 1 1/2 r 1/2 r n Convergence! Ŵ W 2 F N n 1 1/2 := ( ( σ 2 n 1 ) 1/2 r 1/2 O p M 1 K K k=1 1/nk ) 2, r 1/2 := ( 1 K K k=1 rk ) 2
40 What we want to derive: Ŵ W 2 N Proof idea Since Ŵ minimizes the objecfve, EsFmated tensor Obj(Ŵ) Obj(W ) It is not so hard to see: X (ɛ) = F True lowrank tensor 1 2M X(Ŵ W ) 2 2 X (ɛ)/m, Ŵ W + λm Ŵ W S1 ( ) c(n, r) O p M M ɛ i X i i=1
41 EsFmated tensor Proof outline (1/3) True lowrank tensor 1 2M X(Ŵ W ) 2 2 X (ɛ)/m, Ŵ W + λm Ŵ W S1 Inequality 1: upper bound the dot product X (ɛ)/m, Ŵ W O p ( σ 2 N n 1 1/2 M Ŵ W S1 ) (opfmizafon duality / random matrix theory)
42 Proof outline (1/3) EsFmated tensor True lowrank tensor 1 2M X(Ŵ W ) 2 2 ( σ 2 N n 1 1/2 M + λ M ) Ŵ W S 1 Inequality 1: upper bound the dot product X (ɛ)/m, Ŵ W O p ( σ 2 N n 1 1/2 M Ŵ W S1 ) Trade off between OpFmal reg. const and λ M O p ( σ 2 N n 1 1/2 M )
43 Proof outline (2/3) EsFmated True lowrank tensor tensor 1 2M X(Ŵ W ) 2 2 σ 2 N n 1 1/2 M Inequality 2: relate the scha-en 1 norm with the Frobenius norm Ŵ W S1 Ŵ W S1 r 1/2 Ŵ W F (relafon between L1 and L2 norm)
44 EsFmated tensor Proof outline (2/3) True lowrank tensor 1 2M X(Ŵ W ) 2 2 Inequality 2: relate the scha-en 1 norm with the Frobenius norm σ 2 N n 1 1/2 r 1/2 M Ŵ W S1 r 1/2 Ŵ W F (relafon between L1 and L2 norm) Ŵ W F
45 EsFmated tensor Proof outline (3/3) Inequality 3: lower bound the les hand side κ True lowrank tensor 1 2M X(Ŵ W ) 2 2 Ŵ W 2 σ 2 N n 1 1/2 r 1/2 M F 1 M X(Ŵ W ) 2 2 Ŵ W F If #samples (M) #variables (N) c 1 n 1 1/2 r 1/2 (Gordon Slepian Theorem in Gaussian process theory)
46 EsFmated tensor κ Proof outline (3/3) Ŵ W 2 Inequality 3: lower bound the les hand side κ True lowrank tensor F Ŵ W 2 σ 2 N n 1 1/2 r 1/2 M F 1 M X(Ŵ W ) 2 2 Ŵ W F If #samples (M) #variables (N) c 1 n 1 1/2 r 1/2 (Gordon Slepian Theorem in Gaussian process theory)
47 Back to the theorem statement Assume elements of X i are drown iid from standard normal distribufon Moreover Convergence! Ŵ W 2 F N NoFce: #samples (M) #variables (N) c 1 n 1 1/2 r 1/2 r n n 1 1/2 := ( σ 2 n 1 ) 1/2 r 1/2 O p M ( 1 K K k=1 1/nk ) 2, r 1/2 := ( 1 K K k=1 Sample size condifon independent of noise σ 2. Bound RHS proporfonal to σ 2. Threshold behavior in the limit σ 2 0 rk ) 2
48 : /015**/* '!! '!!( 1 Tensor completion results!!"#!"$!"%!"& ' )*+,-./01/21/345* #samples (M) #variables (N) 7*,89.)32,92:**)*;<!"!' 1 '!"&!"%!"$!"# ;/065<1=>1&1?@ ;/65<1=$!1?1>@ AB-.9.C+-./01-/85*+0,5 =./0<>?!2?!2#!@ =./0<>'!!2'!!2?!@! 2!!"#!"$!"%!"& 2 ()*+,-./012*,342553!' 55 '6# 55*55 '6#
49 Including 4 th order tensors ' 2 7*,89.)32,920**:;!"!'!"&!"%!"$!"# Gap between necessity & sufficiency? <./0;=>!2>!2#!? <./0;='!!2'!!2>!? <./0;=>!2>!2#!2'!? <./0;='!!2'!!2#!2'!?! 2!!"#!"$!"%!"& ' ()*+,-./012*,342553!' 55 55*55 '6# '6#
50 9,.:;0+54.;42,,<=!"!# #!")!"(!"& Matrix / tensor completion K = 2 >012=?'!4$!@!"$ >012=?#!!4&!@ >012=?$'!4$!!@! 4!!"#!"$!"%!"&!"'!"( *+,-./01234, !# 77 #8$ 77,77 #8$ 4 7*,89.)32,920**:;!"!' '!"&!"%!"$!"# K 3 <./0;=>!2>!2#!? <./0;='!!2'!!2>!? <./0;=>!2>!2#!2'!? <./0;='!!2'!!2#!2'!?! 2!!"#!"$!"%!"& ' ()*+,-./012*,342553!' 55 55*55 '6# '6# 2
51 Conclusion Many real world problems can be cast into the form of tensor data analysis. Convex opfmizafon is a useful tool also for the analysis of higher order tensors. Proposed a convex tensor decomposifon algorithm with performance guarantee Normalized rank predicts empirical scaling behavior well
52 Issues Why matrix complefon more difficult than tensor complefon? How big the gap between necessity and sufficiency? Random Gaussian design tensor complefon Incoherence (Candes & Recht 09) Spikiness (Negahban et al 10) When only some modes are low rank Scha-en 1 norm is too strong Mixture approach E.g. Mode 1, 4 is low rank but the rest is not (combinatorial problem)
53 T h a n k y o u!
ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis
ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis Lecture 7: Matrix completion Yuejie Chi The Ohio State University Page 1 Reference Guaranteed Minimum-Rank Solutions of Linear
More informationSome tensor decomposition methods for machine learning
Some tensor decomposition methods for machine learning Massimiliano Pontil Istituto Italiano di Tecnologia and University College London 16 August 2016 1 / 36 Outline Problem and motivation Tucker decomposition
More informationTensor Decomposition with Smoothness (ICML2017)
Tensor Decomposition with Smoothness (ICML2017) Masaaki Imaizumi 1 Kohei Hayashi 2,3 1 Institute of Statistical Mathematics 2 National Institute of Advanced Industrial Science and Technology 3 RIKEN Center
More informationBEYOND MATRIX COMPLETION
BEYOND MATRIX COMPLETION ANKUR MOITRA MASSACHUSETTS INSTITUTE OF TECHNOLOGY Based on joint work with Boaz Barak (MSR) Part I: RecommendaIon systems and parially observed matrices THE NETFLIX PROBLEM movies
More informationSpectral k-support Norm Regularization
Spectral k-support Norm Regularization Andrew McDonald Department of Computer Science, UCL (Joint work with Massimiliano Pontil and Dimitris Stamos) 25 March, 2015 1 / 19 Problem: Matrix Completion Goal:
More informationhttps://goo.gl/kfxweg KYOTO UNIVERSITY Statistical Machine Learning Theory Sparsity Hisashi Kashima kashima@i.kyoto-u.ac.jp DEPARTMENT OF INTELLIGENCE SCIENCE AND TECHNOLOGY 1 KYOTO UNIVERSITY Topics:
More informationCollaborative Filtering Matrix Completion Alternating Least Squares
Case Study 4: Collaborative Filtering Collaborative Filtering Matrix Completion Alternating Least Squares Machine Learning for Big Data CSE547/STAT548, University of Washington Sham Kakade May 19, 2016
More informationAdaptive one-bit matrix completion
Adaptive one-bit matrix completion Joseph Salmon Télécom Paristech, Institut Mines-Télécom Joint work with Jean Lafond (Télécom Paristech) Olga Klopp (Crest / MODAL X, Université Paris Ouest) Éric Moulines
More informationSparsity Models. Tong Zhang. Rutgers University. T. Zhang (Rutgers) Sparsity Models 1 / 28
Sparsity Models Tong Zhang Rutgers University T. Zhang (Rutgers) Sparsity Models 1 / 28 Topics Standard sparse regression model algorithms: convex relaxation and greedy algorithm sparse recovery analysis:
More informationComputational Linear Algebra
Computational Linear Algebra PD Dr. rer. nat. habil. Ralf-Peter Mundani Computation in Engineering / BGU Scientific Computing in Computer Science / INF Winter Term 2018/19 Part 6: Some Other Stuff PD Dr.
More informationRank Determination for Low-Rank Data Completion
Journal of Machine Learning Research 18 017) 1-9 Submitted 7/17; Revised 8/17; Published 9/17 Rank Determination for Low-Rank Data Completion Morteza Ashraphijuo Columbia University New York, NY 1007,
More informationCSC 576: Variants of Sparse Learning
CSC 576: Variants of Sparse Learning Ji Liu Department of Computer Science, University of Rochester October 27, 205 Introduction Our previous note basically suggests using l norm to enforce sparsity in
More informationCollaborative Filtering
Case Study 4: Collaborative Filtering Collaborative Filtering Matrix Completion Alternating Least Squares Machine Learning/Statistics for Big Data CSE599C1/STAT592, University of Washington Carlos Guestrin
More informationBEYOND MATRIX COMPLETION
BEYOND MATRIX COMPLETION ANKUR MOITRA MASSACHUSETTS INSTITUTE OF TECHNOLOGY Based on joint work with Boaz Barak (Harvard) Part I: RecommendaJon systems and parjally observed matrices THE NETFLIX PROBLEM
More informationECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference
ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and Inference Low-rank matrix recovery via convex relaxations Yuejie Chi Department of Electrical and Computer Engineering Spring
More informationMatrix Factorizations: A Tale of Two Norms
Matrix Factorizations: A Tale of Two Norms Nati Srebro Toyota Technological Institute Chicago Maximum Margin Matrix Factorization S, Jason Rennie, Tommi Jaakkola (MIT), NIPS 2004 Rank, Trace-Norm and Max-Norm
More informationA Fast Augmented Lagrangian Algorithm for Learning Low-Rank Matrices
A Fast Augmented Lagrangian Algorithm for Learning Low-Rank Matrices Ryota Tomioka 1, Taiji Suzuki 1, Masashi Sugiyama 2, Hisashi Kashima 1 1 The University of Tokyo 2 Tokyo Institute of Technology 2010-06-22
More informationProvable Alternating Minimization Methods for Non-convex Optimization
Provable Alternating Minimization Methods for Non-convex Optimization Prateek Jain Microsoft Research, India Joint work with Praneeth Netrapalli, Sujay Sanghavi, Alekh Agarwal, Animashree Anandkumar, Rashish
More informationLinear dimensionality reduction for data analysis
Linear dimensionality reduction for data analysis Nicolas Gillis Joint work with Robert Luce, François Glineur, Stephen Vavasis, Robert Plemmons, Gabriella Casalino The setup Dimensionality reduction for
More informationLinear Algebra (Review) Volker Tresp 2017
Linear Algebra (Review) Volker Tresp 2017 1 Vectors k is a scalar (a number) c is a column vector. Thus in two dimensions, c = ( c1 c 2 ) (Advanced: More precisely, a vector is defined in a vector space.
More informationA Characterization of Sampling Patterns for Union of Low-Rank Subspaces Retrieval Problem
A Characterization of Sampling Patterns for Union of Low-Rank Subspaces Retrieval Problem Morteza Ashraphijuo Columbia University ashraphijuo@ee.columbia.edu Xiaodong Wang Columbia University wangx@ee.columbia.edu
More informationMatrix Completion: Fundamental Limits and Efficient Algorithms
Matrix Completion: Fundamental Limits and Efficient Algorithms Sewoong Oh PhD Defense Stanford University July 23, 2010 1 / 33 Matrix completion Find the missing entries in a huge data matrix 2 / 33 Example
More information1 Regression with High Dimensional Data
6.883 Learning with Combinatorial Structure ote for Lecture 11 Instructor: Prof. Stefanie Jegelka Scribe: Xuhong Zhang 1 Regression with High Dimensional Data Consider the following regression problem:
More informationLinear Algebra (Review) Volker Tresp 2018
Linear Algebra (Review) Volker Tresp 2018 1 Vectors k, M, N are scalars A one-dimensional array c is a column vector. Thus in two dimensions, ( ) c1 c = c 2 c i is the i-th component of c c T = (c 1, c
More informationMatrix Completion from Fewer Entries
from Fewer Entries Stanford University March 30, 2009 Outline The problem, a look at the data, and some results (slides) 2 Proofs (blackboard) arxiv:090.350 The problem, a look at the data, and some results
More informationStructured matrix factorizations. Example: Eigenfaces
Structured matrix factorizations Example: Eigenfaces An extremely large variety of interesting and important problems in machine learning can be formulated as: Given a matrix, find a matrix and a matrix
More informationRecovering any low-rank matrix, provably
Recovering any low-rank matrix, provably Rachel Ward University of Texas at Austin October, 2014 Joint work with Yudong Chen (U.C. Berkeley), Srinadh Bhojanapalli and Sujay Sanghavi (U.T. Austin) Matrix
More informationTensor Low-Rank Completion and Invariance of the Tucker Core
Tensor Low-Rank Completion and Invariance of the Tucker Core Shuzhong Zhang Department of Industrial & Systems Engineering University of Minnesota zhangs@umn.edu Joint work with Bo JIANG, Shiqian MA, and
More informationLecture Notes 10: Matrix Factorization
Optimization-based data analysis Fall 207 Lecture Notes 0: Matrix Factorization Low-rank models. Rank- model Consider the problem of modeling a quantity y[i, j] that depends on two indices i and j. To
More informationMaximum Margin Matrix Factorization
Maximum Margin Matrix Factorization Nati Srebro Toyota Technological Institute Chicago Joint work with Noga Alon Tel-Aviv Yonatan Amit Hebrew U Alex d Aspremont Princeton Michael Fink Hebrew U Tommi Jaakkola
More informationGI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis. Massimiliano Pontil
GI07/COMPM012: Mathematical Programming and Research Methods (Part 2) 2. Least Squares and Principal Components Analysis Massimiliano Pontil 1 Today s plan SVD and principal component analysis (PCA) Connection
More informationGeneralized Higher-Order Tensor Decomposition via Parallel ADMM
Proceedings of the Twenty-Eighth AAAI Conference on Artificial Intelligence Generalized Higher-Order Tensor Decomposition via Parallel ADMM Fanhua Shang 1, Yuanyuan Liu 2, James Cheng 1 1 Department of
More informationProbabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms
Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms François Caron Department of Statistics, Oxford STATLEARN 2014, Paris April 7, 2014 Joint work with Adrien Todeschini,
More informationLow-rank Matrix Completion with Noisy Observations: a Quantitative Comparison
Low-rank Matrix Completion with Noisy Observations: a Quantitative Comparison Raghunandan H. Keshavan, Andrea Montanari and Sewoong Oh Electrical Engineering and Statistics Department Stanford University,
More informationMatrix completion: Fundamental limits and efficient algorithms. Sewoong Oh Stanford University
Matrix completion: Fundamental limits and efficient algorithms Sewoong Oh Stanford University 1 / 35 Low-rank matrix completion Low-rank Data Matrix Sparse Sampled Matrix Complete the matrix from small
More informationRobust Tensor Completion and its Applications. Michael K. Ng Department of Mathematics Hong Kong Baptist University
Robust Tensor Completion and its Applications Michael K. Ng Department of Mathematics Hong Kong Baptist University mng@math.hkbu.edu.hk Example Low-dimensional Structure Data in many real applications
More informationMatrix Rank Minimization with Applications
Matrix Rank Minimization with Applications Maryam Fazel Haitham Hindi Stephen Boyd Information Systems Lab Electrical Engineering Department Stanford University 8/2001 ACC 01 Outline Rank Minimization
More informationRobust Principal Component Analysis
ELE 538B: Mathematics of High-Dimensional Data Robust Principal Component Analysis Yuxin Chen Princeton University, Fall 2018 Disentangling sparse and low-rank matrices Suppose we are given a matrix M
More informationConditions for Robust Principal Component Analysis
Rose-Hulman Undergraduate Mathematics Journal Volume 12 Issue 2 Article 9 Conditions for Robust Principal Component Analysis Michael Hornstein Stanford University, mdhornstein@gmail.com Follow this and
More informationLow Rank Matrix Completion Formulation and Algorithm
1 2 Low Rank Matrix Completion and Algorithm Jian Zhang Department of Computer Science, ETH Zurich zhangjianthu@gmail.com March 25, 2014 Movie Rating 1 2 Critic A 5 5 Critic B 6 5 Jian 9 8 Kind Guy B 9
More informationUsing SVD to Recommend Movies
Michael Percy University of California, Santa Cruz Last update: December 12, 2009 Last update: December 12, 2009 1 / Outline 1 Introduction 2 Singular Value Decomposition 3 Experiments 4 Conclusion Last
More information15-780: LinearProgramming
15-780: LinearProgramming J. Zico Kolter February 1-3, 2016 1 Outline Introduction Some linear algebra review Linear programming Simplex algorithm Duality and dual simplex 2 Outline Introduction Some linear
More informationRank minimization via the γ 2 norm
Rank minimization via the γ 2 norm Troy Lee Columbia University Adi Shraibman Weizmann Institute Rank Minimization Problem Consider the following problem min X rank(x) A i, X b i for i = 1,..., k Arises
More informationThe convex algebraic geometry of rank minimization
The convex algebraic geometry of rank minimization Pablo A. Parrilo Laboratory for Information and Decision Systems Massachusetts Institute of Technology International Symposium on Mathematical Programming
More informationsparse and low-rank tensor recovery Cubic-Sketching
Sparse and Low-Ran Tensor Recovery via Cubic-Setching Guang Cheng Department of Statistics Purdue University www.science.purdue.edu/bigdata CCAM@Purdue Math Oct. 27, 2017 Joint wor with Botao Hao and Anru
More informationLow-Rank Matrix Recovery
ELE 538B: Mathematics of High-Dimensional Data Low-Rank Matrix Recovery Yuxin Chen Princeton University, Fall 2018 Outline Motivation Problem setup Nuclear norm minimization RIP and low-rank matrix recovery
More informationAn iterative hard thresholding estimator for low rank matrix recovery
An iterative hard thresholding estimator for low rank matrix recovery Alexandra Carpentier - based on a joint work with Arlene K.Y. Kim Statistical Laboratory, Department of Pure Mathematics and Mathematical
More information1-Bit Matrix Completion
1-Bit Matrix Completion Mark A. Davenport School of Electrical and Computer Engineering Georgia Institute of Technology Yaniv Plan Mary Wootters Ewout van den Berg Matrix Completion d When is it possible
More informationAcyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs
Acyclic Semidefinite Approximations of Quadratically Constrained Quadratic Programs Raphael Louca & Eilyan Bitar School of Electrical and Computer Engineering American Control Conference (ACC) Chicago,
More informationLecture 3. Random Fourier measurements
Lecture 3. Random Fourier measurements 1 Sampling from Fourier matrices 2 Law of Large Numbers and its operator-valued versions 3 Frames. Rudelson s Selection Theorem Sampling from Fourier matrices Our
More informationBinary matrix completion
Binary matrix completion Yaniv Plan University of Michigan SAMSI, LDHD workshop, 2013 Joint work with (a) Mark Davenport (b) Ewout van den Berg (c) Mary Wootters Yaniv Plan (U. Mich.) Binary matrix completion
More informationProbabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms
Probabilistic Low-Rank Matrix Completion with Adaptive Spectral Regularization Algorithms Adrien Todeschini Inria Bordeaux JdS 2014, Rennes Aug. 2014 Joint work with François Caron (Univ. Oxford), Marie
More informationAn Extended Frank-Wolfe Method, with Application to Low-Rank Matrix Completion
An Extended Frank-Wolfe Method, with Application to Low-Rank Matrix Completion Robert M. Freund, MIT joint with Paul Grigas (UC Berkeley) and Rahul Mazumder (MIT) CDC, December 2016 1 Outline of Topics
More informationRank, Trace-Norm & Max-Norm
Rank, Trace-Norm & Max-Norm as measures of matrix complexity Nati Srebro University of Toronto Adi Shraibman Hebrew University Matrix Learning users movies 2 1 4 5 5 4? 1 3 3 5 2 4? 5 3? 4 1 3 5 2 1? 4
More informationLecture 20: November 1st
10-725: Optimization Fall 2012 Lecture 20: November 1st Lecturer: Geoff Gordon Scribes: Xiaolong Shen, Alex Beutel Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These notes have not
More informationTensor Analysis. Topics in Data Mining Fall Bruno Ribeiro
Tensor Analysis Topics in Data Mining Fall 2015 Bruno Ribeiro Tensor Basics But First 2 Mining Matrices 3 Singular Value Decomposition (SVD) } X(i,j) = value of user i for property j i 2 j 5 X(Alice, cholesterol)
More information1-Bit Matrix Completion
1-Bit Matrix Completion Mark A. Davenport School of Electrical and Computer Engineering Georgia Institute of Technology Yaniv Plan Mary Wootters Ewout van den Berg Matrix Completion d When is it possible
More informationFactor Matrix Trace Norm Minimization for Low-Rank Tensor Completion
Factor Matrix Trace Norm Minimization for Low-Rank Tensor Completion Yuanyuan Liu Fanhua Shang Hong Cheng James Cheng Hanghang Tong Abstract Most existing low-n-rank imization algorithms for tensor completion
More informationTensor Robust Principal Component Analysis: Exact Recovery of Corrupted Low-Rank Tensors via Convex Optimization
Tensor Robust Principal Component Analysis: Exact Recovery of Corrupted Low-Rank Tensors via Convex Optimization Canyi Lu 1, Jiashi Feng 1, Yudong Chen, Wei Liu 3, Zhouchen Lin 4,5,, Shuicheng Yan 6,1
More informationMatrix Completion for Structured Observations
Matrix Completion for Structured Observations Denali Molitor Department of Mathematics University of California, Los ngeles Los ngeles, C 90095, US Email: dmolitor@math.ucla.edu Deanna Needell Department
More informationCS264: Beyond Worst-Case Analysis Lecture #15: Topic Modeling and Nonnegative Matrix Factorization
CS264: Beyond Worst-Case Analysis Lecture #15: Topic Modeling and Nonnegative Matrix Factorization Tim Roughgarden February 28, 2017 1 Preamble This lecture fulfills a promise made back in Lecture #1,
More informationBasic Concepts in Linear Algebra
Basic Concepts in Linear Algebra Grady B Wright Department of Mathematics Boise State University February 2, 2015 Grady B Wright Linear Algebra Basics February 2, 2015 1 / 39 Numerical Linear Algebra Linear
More informationInformation-Theoretic Limits of Matrix Completion
Information-Theoretic Limits of Matrix Completion Erwin Riegler, David Stotz, and Helmut Bölcskei Dept. IT & EE, ETH Zurich, Switzerland Email: {eriegler, dstotz, boelcskei}@nari.ee.ethz.ch Abstract We
More informationCompressed Sensing and Robust Recovery of Low Rank Matrices
Compressed Sensing and Robust Recovery of Low Rank Matrices M. Fazel, E. Candès, B. Recht, P. Parrilo Electrical Engineering, University of Washington Applied and Computational Mathematics Dept., Caltech
More informationCHAPTER 11. A Revision. 1. The Computers and Numbers therein
CHAPTER A Revision. The Computers and Numbers therein Traditional computer science begins with a finite alphabet. By stringing elements of the alphabet one after another, one obtains strings. A set of
More informationLecture 8: Linear Algebra Background
CSE 521: Design and Analysis of Algorithms I Winter 2017 Lecture 8: Linear Algebra Background Lecturer: Shayan Oveis Gharan 2/1/2017 Scribe: Swati Padmanabhan Disclaimer: These notes have not been subjected
More informationApproximate Low-Rank Tensor Learning
Approximate Low-Rank Tensor Learning Yaoliang Yu Dept of Machine Learning Carnegie Mellon University yaoliang@cs.cmu.edu Hao Cheng Dept of Electric Engineering University of Washington kelvinwsch@gmail.com
More informationReview of Basic Concepts in Linear Algebra
Review of Basic Concepts in Linear Algebra Grady B Wright Department of Mathematics Boise State University September 7, 2017 Math 565 Linear Algebra Review September 7, 2017 1 / 40 Numerical Linear Algebra
More informationHigh Dimensional Covariance and Precision Matrix Estimation
High Dimensional Covariance and Precision Matrix Estimation Wei Wang Washington University in St. Louis Thursday 23 rd February, 2017 Wei Wang (Washington University in St. Louis) High Dimensional Covariance
More informationInformation Retrieval
Introduction to Information CS276: Information and Web Search Christopher Manning and Pandu Nayak Lecture 13: Latent Semantic Indexing Ch. 18 Today s topic Latent Semantic Indexing Term-document matrices
More informationCS598 Machine Learning in Computational Biology (Lecture 5: Matrix - part 2) Professor Jian Peng Teaching Assistant: Rongda Zhu
CS598 Machine Learning in Computational Biology (Lecture 5: Matrix - part 2) Professor Jian Peng Teaching Assistant: Rongda Zhu Feature engineering is hard 1. Extract informative features from domain knowledge
More informationarxiv: v1 [math.oc] 11 Jun 2009
RANK-SPARSITY INCOHERENCE FOR MATRIX DECOMPOSITION VENKAT CHANDRASEKARAN, SUJAY SANGHAVI, PABLO A. PARRILO, S. WILLSKY AND ALAN arxiv:0906.2220v1 [math.oc] 11 Jun 2009 Abstract. Suppose we are given a
More informationSparse and Low-Rank Matrix Decompositions
Forty-Seventh Annual Allerton Conference Allerton House, UIUC, Illinois, USA September 30 - October 2, 2009 Sparse and Low-Rank Matrix Decompositions Venkat Chandrasekaran, Sujay Sanghavi, Pablo A. Parrilo,
More informationAn Introduction to Sparse Approximation
An Introduction to Sparse Approximation Anna C. Gilbert Department of Mathematics University of Michigan Basic image/signal/data compression: transform coding Approximate signals sparsely Compress images,
More informationThe Stability of Low-Rank Matrix Reconstruction: a Constrained Singular Value Perspective
Forty-Eighth Annual Allerton Conference Allerton House UIUC Illinois USA September 9 - October 1 010 The Stability of Low-Rank Matrix Reconstruction: a Constrained Singular Value Perspective Gongguo Tang
More information1-Bit Matrix Completion
1-Bit Matrix Completion Mark A. Davenport School of Electrical and Computer Engineering Georgia Institute of Technology Yaniv Plan Mary Wootters Ewout van den Berg Matrix Completion d When is it possible
More informationIntroduction to Tensors. 8 May 2014
Introduction to Tensors 8 May 2014 Introduction to Tensors What is a tensor? Basic Operations CP Decompositions and Tensor Rank Matricization and Computing the CP Dear Tullio,! I admire the elegance of
More informationConcentration-Based Guarantees for Low-Rank Matrix Reconstruction
JMLR: Workshop and Conference Proceedings 9 20 35 339 24th Annual Conference on Learning Theory Concentration-Based Guarantees for Low-Rank Matrix Reconstruction Rina Foygel Department of Statistics, University
More informationNon-Convex Projected Gradient Descent for Generalized Low-Rank Tensor Regression
Non-Convex Projected Gradient Descent for Generalized Low-Rank Tensor Regression Han Chen, Garvesh Raskutti and Ming Yuan University of Wisconsin-Madison Morgridge Institute for Research and Department
More informationLarge Scale Data Analysis Using Deep Learning
Large Scale Data Analysis Using Deep Learning Linear Algebra U Kang Seoul National University U Kang 1 In This Lecture Overview of linear algebra (but, not a comprehensive survey) Focused on the subset
More informationAndriy Mnih and Ruslan Salakhutdinov
MATRIX FACTORIZATION METHODS FOR COLLABORATIVE FILTERING Andriy Mnih and Ruslan Salakhutdinov University of Toronto, Machine Learning Group 1 What is collaborative filtering? The goal of collaborative
More informationCourse Notes for EE227C (Spring 2018): Convex Optimization and Approximation
Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Instructor: Moritz Hardt Email: hardt+ee227c@berkeley.edu Graduate Instructor: Max Simchowitz Email: msimchow+ee227c@berkeley.edu
More informationFaloutsos, Tong ICDE, 2009
Large Graph Mining: Patterns, Tools and Case Studies Christos Faloutsos Hanghang Tong CMU Copyright: Faloutsos, Tong (29) 2-1 Outline Part 1: Patterns Part 2: Matrix and Tensor Tools Part 3: Proximity
More informationOn Tensor Train Rank Minimization: Statistical Efficiency and Scalable Algorithm
On Tensor Train Ran Minimization: Statistical Efficiency and Scalable Algorithm Masaai Imaizumi Institute of Statistical Mathematics RIKEN Center for Advanced Intelligence Project imaizumi@ism.ac.jp Taanori
More informationMatrix and Tensor Factorization from a Machine Learning Perspective
Matrix and Tensor Factorization from a Machine Learning Perspective Christoph Freudenthaler Information Systems and Machine Learning Lab, University of Hildesheim Research Seminar, Vienna University of
More informationRecovery of Simultaneously Structured Models using Convex Optimization
Recovery of Simultaneously Structured Models using Convex Optimization Maryam Fazel University of Washington Joint work with: Amin Jalali (UW), Samet Oymak and Babak Hassibi (Caltech) Yonina Eldar (Technion)
More informationHigh-dimensional Statistics
High-dimensional Statistics Pradeep Ravikumar UT Austin Outline 1. High Dimensional Data : Large p, small n 2. Sparsity 3. Group Sparsity 4. Low Rank 1 Curse of Dimensionality Statistical Learning: Given
More informationHigh-Rank Matrix Completion and Subspace Clustering with Missing Data
High-Rank Matrix Completion and Subspace Clustering with Missing Data Authors: Brian Eriksson, Laura Balzano and Robert Nowak Presentation by Wang Yuxiang 1 About the authors
More informationKnowledge Discovery and Data Mining 1 (VO) ( )
Knowledge Discovery and Data Mining 1 (VO) (707.003) Review of Linear Algebra Denis Helic KTI, TU Graz Oct 9, 2014 Denis Helic (KTI, TU Graz) KDDM1 Oct 9, 2014 1 / 74 Big picture: KDDM Probability Theory
More informationCourse Notes for EE227C (Spring 2018): Convex Optimization and Approximation
Course Notes for EE7C (Spring 08): Convex Optimization and Approximation Instructor: Moritz Hardt Email: hardt+ee7c@berkeley.edu Graduate Instructor: Max Simchowitz Email: msimchow+ee7c@berkeley.edu October
More informationSTA141C: Big Data & High Performance Statistical Computing
STA141C: Big Data & High Performance Statistical Computing Numerical Linear Algebra Background Cho-Jui Hsieh UC Davis May 15, 2018 Linear Algebra Background Vectors A vector has a direction and a magnitude
More informationTensor Decompositions for Machine Learning. G. Roeder 1. UBC Machine Learning Reading Group, June University of British Columbia
Network Feature s Decompositions for Machine Learning 1 1 Department of Computer Science University of British Columbia UBC Machine Learning Group, June 15 2016 1/30 Contact information Network Feature
More informationSPARSE signal representations have gained popularity in recent
6958 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 57, NO. 10, OCTOBER 2011 Blind Compressed Sensing Sivan Gleichman and Yonina C. Eldar, Senior Member, IEEE Abstract The fundamental principle underlying
More informationTensor Decompositions and Applications
Tamara G. Kolda and Brett W. Bader Part I September 22, 2015 What is tensor? A N-th order tensor is an element of the tensor product of N vector spaces, each of which has its own coordinate system. a =
More informationLearning with matrix and tensor based models using low-rank penalties
Learning with matrix and tensor based models using low-rank penalties Johan Suykens KU Leuven, ESAT-SCD/SISTA Kasteelpark Arenberg 10 B-3001 Leuven (Heverlee), Belgium Email: johan.suykens@esat.kuleuven.be
More informationSignal Recovery from Permuted Observations
EE381V Course Project Signal Recovery from Permuted Observations 1 Problem Shanshan Wu (sw33323) May 8th, 2015 We start with the following problem: let s R n be an unknown n-dimensional real-valued signal,
More informationGeometric Modeling Summer Semester 2010 Mathematical Tools (1)
Geometric Modeling Summer Semester 2010 Mathematical Tools (1) Recap: Linear Algebra Today... Topics: Mathematical Background Linear algebra Analysis & differential geometry Numerical techniques Geometric
More informationNon-Convex Projected Gradient Descent for Generalized Low-Rank Tensor Regression
Journal of Machine Learning Research 20 (2019) 1-37 Submitted 12/16; Revised 1/18; Published 2/19 Non-Convex Projected Gradient Descent for Generalized Low-Rank Tensor Regression Han Chen Garvesh Raskutti
More informationExact Low-rank Matrix Recovery via Nonconvex M p -Minimization
Exact Low-rank Matrix Recovery via Nonconvex M p -Minimization Lingchen Kong and Naihua Xiu Department of Applied Mathematics, Beijing Jiaotong University, Beijing, 100044, People s Republic of China E-mail:
More informationOptimisation Combinatoire et Convexe.
Optimisation Combinatoire et Convexe. Low complexity models, l 1 penalties. A. d Aspremont. M1 ENS. 1/36 Today Sparsity, low complexity models. l 1 -recovery results: three approaches. Extensions: matrix
More information