EECS 275 Matrix Computation

Similar documents
Subspace sampling and relative-error matrix approximation

Fast Monte Carlo Algorithms for Matrix Operations & Massive Data Set Analysis

EECS 275 Matrix Computation

Relative-Error CUR Matrix Decompositions

RandNLA: Randomized Numerical Linear Algebra

DATA MINING LECTURE 8. Dimensionality Reduction PCA -- SVD

RandNLA: Randomization in Numerical Linear Algebra

CS60021: Scalable Data Mining. Dimensionality Reduction

Randomized algorithms for the approximation of matrices

EECS 275 Matrix Computation

EECS 275 Matrix Computation

Jeffrey D. Ullman Stanford University

Lecture 14: SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Lecturer: Sanjeev Arora

Lecture 9: Numerical Linear Algebra Primer (February 11st)

Methods for sparse analysis of high-dimensional data, II

SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices)

Lecture 18 Nov 3rd, 2015

EECS 275 Matrix Computation

A fast randomized algorithm for approximating an SVD of a matrix

An Experimental Evaluation of a Monte-Carlo Algorithm for Singular Value Decomposition

EECS 275 Matrix Computation

Clustering Large Graphs via the Singular Value Decomposition

Randomized Algorithms in Linear Algebra and Applications in Data Analysis

Sketching as a Tool for Numerical Linear Algebra

Methods for sparse analysis of high-dimensional data, II

Principal Component Analysis

Orthogonal tensor decomposition

FAST MONTE CARLO ALGORITHMS FOR MATRICES III: COMPUTING A COMPRESSED APPROXIMATE MATRIX DECOMPOSITION

sublinear time low-rank approximation of positive semidefinite matrices Cameron Musco (MIT) and David P. Woodru (CMU)

to be more efficient on enormous scale, in a stream, or in distributed settings.

A randomized algorithm for approximating the SVD of a matrix

Mining of Massive Datasets Jure Leskovec, AnandRajaraman, Jeff Ullman Stanford University

Data Mining Lecture 4: Covariance, EVD, PCA & SVD

Large Scale Data Analysis Using Deep Learning

EECS 275 Matrix Computation

Total least squares. Gérard MEURANT. October, 2008

Randomized Numerical Linear Algebra: Review and Progresses

Fast low rank approximations of matrices and tensors

Numerical Methods I Singular Value Decomposition

Pseudoinverse & Moore-Penrose Conditions


Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin

Deep Learning Book Notes Chapter 2: Linear Algebra

Linear Algebra (Review) Volker Tresp 2017

Matrix Decomposition in Privacy-Preserving Data Mining JUN ZHANG DEPARTMENT OF COMPUTER SCIENCE UNIVERSITY OF KENTUCKY

Principal components analysis COMS 4771

Matrix Approximations

COMP6237 Data Mining Covariance, EVD, PCA & SVD. Jonathon Hare

Parallel Numerical Algorithms

14 Singular Value Decomposition

Lecture 4. CP and KSVD Representations. Charles F. Van Loan

Computational Methods. Eigenvalues and Singular Values

Low Rank Matrix Approximation

Sketching as a Tool for Numerical Linear Algebra

Lecture Notes 5: Multiresolution Analysis

1 Singular Value Decomposition and Principal Component

Linear Algebra (Review) Volker Tresp 2018

Linear Algebra and Eigenproblems

Lecture 4. Tensor-Related Singular Value Decompositions. Charles F. Van Loan

15 Singular Value Decomposition

Latent Semantic Analysis. Hongning Wang

FAST MONTE CARLO ALGORITHMS FOR MATRICES II: COMPUTING A LOW-RANK APPROXIMATION TO A MATRIX

Approximate Spectral Clustering via Randomized Sketching

SVD, PCA & Preprocessing

EE 381V: Large Scale Optimization Fall Lecture 24 April 11

Parallel Singular Value Decomposition. Jiaxing Tan

Non-square diagonal matrices

L26: Advanced dimensionality reduction

Randomized Algorithms

Randomized algorithms in numerical linear algebra

Linear Algebra Review. Vectors

Matrices, Vector Spaces, and Information Retrieval

arxiv: v1 [cs.ds] 16 Aug 2016

A fast randomized algorithm for overdetermined linear least-squares regression

Robust Principal Component Analysis

LECTURE 16: PCA AND SVD

AMS526: Numerical Analysis I (Numerical Linear Algebra for Computational and Data Sciences)

Statistical Pattern Recognition

Scientific Computing

MS&E 318 (CME 338) Large-Scale Numerical Optimization

Singular Value Decomposition and its. SVD and its Applications in Computer Vision

Compressed Sensing and Robust Recovery of Low Rank Matrices

Structure in Data. A major objective in data analysis is to identify interesting features or structure in the data.

PCA and admixture models

Linear Algebra for Machine Learning. Sargur N. Srihari

Assignment 3. Latent Semantic Indexing

MIT 9.520/6.860, Fall 2017 Statistical Learning Theory and Applications. Class 19: Data Representation by Design

a Short Introduction

Focus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations.

AMS526: Numerical Analysis I (Numerical Linear Algebra)

EECS 275 Matrix Computation

Numerical Methods I Non-Square and Sparse Linear Systems

STA141C: Big Data & High Performance Statistical Computing

Properties of Matrices and Operations on Matrices

Applications of Randomized Methods for Decomposing and Simulating from Large Covariance Matrices

Latent Semantic Analysis. Hongning Wang

MANY scientific computations, signal processing, data analysis and machine learning applications lead to large dimensional

Maths for Signals and Systems Linear Algebra in Engineering

dimensionality reduction for k-means and low rank approximation

Transcription:

EECS 275 Matrix Computation Ming-Hsuan Yang Electrical Engineering and Computer Science University of California at Merced Merced, CA 95344 http://faculty.ucmerced.edu/mhyang Lecture 22 1 / 21

Overview Algorithms for Modern Massive Data Sets (MMDS): Explore algorithms for modeling and analyzing massive, high dimensional and nonlinear structured data Bring together computer scientists, computational and applied mathematicians, and practitioners Tools: numerical linear algebra, kernel methods, multilinear algebra, randomized algorithms, optimization, differential geometry, geometry and topology, etc. Organized by Gene Golub et al. in 2006, 2008 and 2010 Slides available at mmds.stanford.edu 2 / 21

Topics Low rank matrix approximation: theory, sampling algorithms Nearest neighbor algorithms and approximation Spectral graph theory and applications Non-negative matrix factorization Kernel methods Algebraic topology and analysis of high dimensional data Higher order statistics, tensors and approximations 3 / 21

Matrix factorization and applications Treat each data point as a vector 2D image 1D vector of intensity values 3D models 1D vector of 3D coordinates Document 1D vector of term frequency Massive data set High dimensional data Find a low dimensional representation using eigendecomposition See O Leary s slides 4 / 21

Low rank matrix approximation SVD is great but computationally expensive for large scale problems Efficient randomized algorithms for low rank approximation with guaranteed error bounds CX algorithm [Drineas and Kanna FOCS 01]: randomly pick k columns A m n C m k X k n CUR algorithm [Drineas and Kannan SODA 03]: randomly pick c columns and r rows A m n C m c U c r R r n Element-wise sampling [Achiloptas and McSherry STOC 01] { Aij /p A m n S m n, S ij = ij, with probability p ij 0, with probability 1 p ij See Kannan s slides and Drineas slides. 5 / 21

Approximating matrix multiplication Given an m by n matrix A and an n by p matrix B, AB = n A (i) B (i) }{{} IR m p i=1 where A (i) are the i-th column of A and B (i) is the i-th row of B Each term is a rank-one matrix Random sampling algorithm fix a set of probabilities pi i = 1,..., n summing up to 1 for t = 1 up to s, set j t = i where Pr(j t = i) = p i (pick s terms of the sum with replacement w.r.t. p i ) approximate AB by the sum of s terms, after scaling AB 1 s 1 A (jt) B s p (jt) t=1 jt }{{} IR m p 6 / 21

Approximating matrix multiplication (cont d) In matrix notation A m n B n p C m s R s p Create C and R i.i.d. trials with replacement For t = 1 up to s, pick a column A (jt) and a row B (jt) with probability Pr(j t = i) = A (i) 2 B (i) 2 n i=1 A(i) 2 B (i) 2 Include A (jt) /(sp jt ) 1/2 as a column of C and B (jt)/(sp jt ) 1/2 as a row of R 7 / 21

Approximating matrix multiplication (cont d) The input matrices are given in sparse unordered representation, e.g., their non-zero entries are presented as triples (i, j, A ij ) in any order The expectation of CR (element-wise) is AB The nonuniform sampling minimizes the variance of this estimator Easy to implement the sampling algorithm in two phases If the matrices are dense the algorithm runs in O(smp) time instead of O(nmp) time Require only O(sm + sp) memory space Does not tamper with the sparsity of the matrices For the above algorithm E( AB CR 2,F ) 1 A F B F s With probability at least 1 δ AB CR 2,F O(log(1/δ)) A F B F s 8 / 21

Special case: B = A If B = A, then the sampling probabilities are Pr(picking i) = A (i) 2 2 n i=1 A(i) 2 2 Also R = C, and the error bounds are = A(i) 2 2 A 2 F E( AA CC 2,F ) 1 s A 2 F Improvement for the spectral norm bound for the special case E( AA CC 2 ) 4 log s s A F A 2 The sampling procedure is slightly different; s columns/rows are kept in expectation, i.e., column i is picked with probability Pr(picking i) = min(1, s A(i) 2 2 A 2 ) F 9 / 21

Approximating SVD in O(n) time The complexity of computing SVD of a m by n matrix A is O(min(mn 2, m 2 n)) (e.g., using Golub-Kahan algorithm) The top few singular vectors/values can be approximated faster using Lanczos/Arnoldi methods Let A k be rank k approximation of A A k is a matrix of rank k such that A A k 2,F is minimized over all rank k matrices Approximate SVD in linear time O(m + n) sample c columns from A and rescale to form the m c matrix C compute the m k matrix H k of the top k left singular vectors C Structural theorem: For any probabilities and number of columns A H k H k A 2 2,F A A k 2 2,F + 2 k AA CC F Algorithmic theorem: If p i = A (i) 2 2 / A 2 F and c 4η2 k/ε 2, then A H k H k A 2 2,F A A k 2 2,F + ε A 2 F 10 / 21

Example of randomized SVD Compute the top k left singular vectors of matrix C and restore them in the 512-by-k matrix H k Original matrix A After sampling columns C H k H k A 11 / 21

Potential problems with SVD Structure in the data is not respected by mathematical operations on the data: Reification: maximum variance directions Interpretability: what does a linear combination of 6000 genes mean? Sparsity: destroyed by orthogonalization Non-negativity: is a convex but not linear algebraic notion Does there exist better low-rank matrix approximation? better structure properties for certain applications better at respecting relevant structure better for Interpretability and informing intuition 12 / 21

CX matrix decomposition Goal: Find A m n C m c X c n so that A CX small in some norm One way to approach this is min X IR c n A CX F = A C(C A) F where C is the pseudoinverse of C SVD of A: A k = U k Σ k Vk where A k is of m n, U k of m k, Σ of k k, and Vk of k n Subspace sampling: V k is an orthogonal matrix containing the top k left singular vectors of A The columns of V k are orthonormal vectors, but the rows of V k, denoted by (V k ) (i) are not orthonormal vectors Subspace sampling in O(SVD k (A)) time i = 1, 2,..., n, p i = (V k) (i) 2 2 k 13 / 21

Relative error CX Relative error CX decomposition compute the probabilities p i for each i = 1, 2,..., n, pick the i-th column of A with probability min(1, cp i ) let C be the matrix containing the sampled columns Theorem For any k, let A k be the best rank k approximation to A. In O(SVD k (A)) we can compute p i such that if c = O(k log k/ε 2 ) then with probability at least 1 δ min X IR c n A CX F = A CC A F (1 + ε) A A k F 14 / 21

CUR decomposition Goal: Find A m n C m c U c r R r n so that A CUR is small in some norm Why: After making two passes over A, one can compute provably C, U, and R and store them (sketch) instead of A of O(m + n) vs. O(mn) SVD of A m n = U m p Σ p p V p n where ran(a) = p Exact computation of the SVD takes O(min(mn 2, m 2 n)) and the top k left/right singular vectors/values can be computed from Lanczos/Arnoldi methods Rank k approximation A k = U k Σ k k Vk matrix with top k singular values of A where Σ k k is a diagonal Note that the columns of U k are linear combinations of all columns of A, and the rows of V k are linear combinations of all rows of A 15 / 21

The CUR decomposition Sampling columns for C: use CX algorithm to sample columns of A C has c columns in expectation C = U C Σ C V C m c m ρ ρ ρ ρ c UC is the orthogonal matrix containing the left singular vectors of C and ρ is the rank of C Let (U C ) (i) denote the i-th row of U Sampling rows for R: subspace sampling in O(c 2 m) time with probability q i i = 1, 2,..., m q i = (U C ) (i) 2 2 ρ R has r rows in expectation Compute U: Let W be the intersection of C and R Let U be a rescaled pseudoinverse of W 16 / 21

The CUR decomposition (cont d) Put together A CUR A C ([ D W ] D) m n m c r c r n where D is a diagonal rescaling matrix and U = (DW ) D R Theorem Given C, in O(c 2 m) time, one can compute q i such that A CUR F (1 + ε) A C(C A) F holds with probability at least 1 δ if r = O(c log c/ε 2 ) rows 17 / 21

Relative error CUR Theorem For any k, it takes O(SVD k (A)) time to construct C, U, and R such that A CUR F (1 + ε) A U k Σ k Vk F = (1 + ε) A A k F holds with probability at least 1 δ by picking O(k log k log(1/δ)/ε 2 ) columns, and O(k log 2 k log(1/δ)/ε 6 ) rows where O(SVD k (A)) is the time to compute the top k top left/right singular values Applications: Genomic microarray data, time-resolved fmri data, sequence and mutational data, hyperspectral color cancer data For small k, in O(SVD k (A)) time we can construct C, U, and R s.t. A CUR F (1 + 0.001) A A k F by typically at most k + 5 columns and at most k + 5 rows 18 / 21

Element-wise sampling Main idea: to approximate matrix A, keep a few elements of the matrix (instead of sampling rows or columns) and zero out the remaining elements compute a rank k approximation to this sparse matrix (using Lanczos methods) Create the matrix S from A such that { A ij /p ij with probability p ij S ij = 0 with probability 1 p ij It can be shown that A S 2 is bounded and the singular values of S and A are close Under additional assumptions the top k left singular vectors of S span a subspace that is close to the subspace spanned by the top k left singular vectors of A 19 / 21

Element-wise sampling (cont d) Approximating singular values fast: zero out a large number of elements of A, and scale the remaining ones appropriately compute the singular values of the resulting sparse matrix using iterative methods good choice for p ij = sa2 ij where s denotes the expected number of i,j A2 ij elements that we seek to keep in S note each element is kept or discarded independently of the others Similar ideas that has been used to explain the success of Latent Semantic Indexing design recommendation systems speed up kernel computation 20 / 21

Element-wise sampling vs. row/column sampling Row/column sampling preserves subspace/structural properties of the matrices Element-wise sampling explains how adding noise and/or quantizing the elements of a matrix perturbs its singular values/vectors These two techniques should be complementary These two techniques have similar error bounds Element-wise sampling can be carried out in one pass Running time of element-wise sampling depends on the speed of iterative methods 21 / 21