Dimensionality reduction: Johnson-Lindenstrauss lemma for structured random matrices

Size: px
Start display at page:

Download "Dimensionality reduction: Johnson-Lindenstrauss lemma for structured random matrices"

Transcription

1 Dimensionality reduction: Johnson-Lindenstrauss lemma for structured random matrices Jan Vybíral Austrian Academy of Sciences RICAM, Linz, Austria January 2011 MPI Leipzig, Germany

2 joint work with Aicke Hinrichs (University of Jena, Germany) Massimo Fornasier (RICAM, Linz) Jan Haškovec (RICAM, Linz) FWF-START project: Sparse Approximation and Optimization in High Dimensions led by Massimo Fornasier

3 Outline Johnson-Lindenstrauss lemma Classical proof Variants and improvements Circulant matrices Decoupling vs. Fourier transform Applications - Approximate nearest neighbours - Dynamical systems

4 Johnson-Lindenstrauss lemma Let ε (0, 1 2 ), x 1,...,x n R d...arbitrary points, k = O(ε 2 log n), i.e. k Cε 2 log n. There exists a (linear) mapping f : R d R k such that (1 ε) x i x j 2 2 f (x i) f (x j ) 2 2 (1 + ε) x i x j 2 2 for all i,j {1,...,n}. Here 2 stands for the Euclidean norm in R d or R k, respectively.

5 Typical proof A R k d - k d matrices, P - probability measure on A For each y, y 2 = 1: concentration of measure Choosing ( ) P(A A : Ay 2 2 > 1 + ε) exp( ckε2 ), ( ) P(A A : Ay 2 2 < 1 ε) exp( ckε 2 ). exp( ckε 2 ) 1 n 2, the probability of failure (union bound) is smaller then ( ) n n 2 = 1 1 n < 1. Hence, the probability of success is positive!

6 The condition exp( ckε 2 ) 1 n 2 leads to and ckε 2 2log n k 2 c ε 2 log n, i.e. C = 2 c. By increasing C(= 3/c), we may achieve, that such a mapping becomes typical, i.e. occurs with probability at least 1 1/n.

7 Proof of ( ) P...rotational invariant U R d d unitary; X, Y A: P(XU) = P(Y) X := {A : Ay 2 2 > 1 + ε}, Y := {B : Be > 1 + ε}, Ue 1 = y a 1,1... a 1,d A = 1 a 2,1... a 2,d k..... a k,1... a k,d a i,j - indep. Gaussian variables (P is a tensor product) k Be = 1 k i=1 b 2 i,1

8 ( k ) P σi 2 > k(1 + ε) exp( ckε 2 )? i=1 ( k ) ( P σi 2 > k(1 + ε) = P λ i=1 = P ( E exp exp ( ( λ λ ) k σi 2 kλ(1 + ε) > 0 i=1 ) ) k σi 2 kλ(1 + ε) > 1 i=1 ) k σi 2 kλ(1 + ε) i=1 ( 1 = e kλ(1+ε) 2π = e kλ(1+ε) (1 2λ) k/2 R ) k e λt2 e t2 /2 dt

9 Optimization of λ: 0 < λ = 1/2(1 1 1+ε ) < 1/2 leads to e kε/2 (1 + ε) k/2 exp( k/2(ε 2 /2 ε 3 /3))...for 0 < ε < 1/2... exp( ckε 2 )

10 Classical proof W. B. Johnson and J. Lindenstrauss, Extensions of Lipschitz mappings into a Hilbert space. Contem. Math., 26: , 1984 Projection onto a random k-dimensional subspace satisfies the desired property with positive probability advantages: geometrical proof disadvantages: measure on the set of all k-dimensional subspaces evaluating f (x) involves orthonormalisation time consuming

11 Variants and improvements Elementary proof: S. Dasgupta and A. Gupta An elementary proof of a theorem of Johnson and Lindenstrauss. Random. Struct. Algorithms, 22:60-65, Improvements motivated by applications: Good running times of f (x) Small randomness used Small memory space used...others...

12 D. Achlioptas, Database-friendly random projections: Johnson-Lindenstrauss with binary coins. J. Comput. Syst. Sci., 66(4): , f realized by a k d matrix, where each entry is generated independently at random: Gaussian or Bernoulli (or similar) variables. Running time: k d Randomness: k d Memory space k d

13 N. Ailon and B. Chazelle, Approximate nearest neighbors and the fast Johnson-Lindenstrauss transform. In Proc. 38th Annual ACM Symposium on Theory of Computing, f (x) = PHDx, where P is a k d matrix, where each component is generated independently at random P i,j = N(0,1) with probability { q = min Θ P i,j = 0 with probability 1 q, ( log 2 n d ) },1 H is the d d normalized Hadamard matrix, D is a random d d diagonal matrix, with each D i,i drawn independently from { 1,1} with probability 1/2.

14 Running time: With high probability, f (x) may be calculated in time O(d log d + qdk) Randomness: O(d + k log 2 n) Memory space: with high probability O(d +kdq) = O(d +k log 2 n) Not easy to implement Other variants and improvements...

15 Connection to compressed sensing - RIP We say, that A R k d satisfies the Restricted Isometry Property of order s, if there exists δ s (0,1), such that (1 δ s ) x 2 2 Ax 2 2 (1 + δ s) x 2 2 holds for all x R d with x 0 := #{j = 1,...,d : x j 0} s. The aim is to find matrices with small δ s for large s.

16 R. Baraniuk, M. Davenport, R. DeVore and M. Wakin A simple proof of the Restricted Isometry Property for Random Matrices, Constructive Approximation, If P ω ( Aω x 2 2 (1 + ε) x 2 2) exp( nc(ε)) (and the same for ), then A ω satisfies RIP of order s and δ s for s small enough with probability depending on s and δ s. Every distribution that yields J-L transforms, yields also RIP-matrices.

17 Reverse direction provided recently by F. Krahmer and R. Ward: New and improved Johnson-Lindenstrauss embeddings via the Restricted Isometry Property If Φ R k d satisfies RIP of order s (large enough), then ΦD ξ satisfies J-L Lemma. Every distribution that yields RIP matrices, yields also J-L-transforms. Here, D ξ R d d is a diagonal matrix with indep. Bernoulli variables on the diagonal.

18 Circulant matrices a = (a 0,...,a d 1 ) be i.i.d. random variables a 0 a 1 a 2... a d 1 a d 1 a 0 a 1... a d 2 M a,k = a d 2 a d 1 a 0... a d 3 R k d a d k+1 a d k+2 a d k+3... a d k Is it possible to take f (x) = 1 k M a,k x? Or f (x) = 1 k M a,k D κ x?

19 Decoupling vs. Fourier transform Yes! With k = O(ε 2 log 3 n) - decoupling techniques Yes! With k = O(ε 2 log 2 n) - Fourier-analytic methods The improvement to O(ε 2 log n) is still open... promising numerical experiments advantages: running time O(d log d) - using FFT randomness used 2d instead of (k + 1)d easy to implement: FFT is a part of every software package disadvantage: up to now - bigger k

20 log 3 n: Let x 1,...,x n be arbitrary points in R d, ε (0, 1 2 ), k = O(ε 2 log 3 n), a = (a 0,...,a d 1 ) be independent Bernoulli variables or independent normally distributed variables, M a,k and D κ be as above and f (x) = 1 k M a,k D κ x. Then with probability at least 2/3 the following holds (1 ε) x i x j 2 2 f (x i ) f (x j ) 2 2 (1+ε) x i x j 2 2, i,j = 1,...,n.

21 Strategy of the proof of log 3 n -decoupling the dependence Concentration inequalities for every fixed x with x 2 = 1: P a,κ ( M a,k D κ x 2 2 (1 + ε)k ) exp( c(kε 2 ) 1/3 ) and P a,κ ( M a,k D κ x 2 2 (1 ε)k ) exp( c(kε 2 ) 1/3 ). Then union bound over all n(n 1)/2 pairs of points. The bound on k is given by 2 n(n 1) 2 exp( c(kε 2 ) 1/3 ) < 1.

22 Separation of the diagonal and the off-diagonal term k 1 M a,k D κ x 2 2 = d 1 k 1 I = ai 2 i=0 j=0 x 2 j+i } {{ } diagonal j=0 ( d 1 ) 2 a i κ j+i x j+i = I + II i=0 k 1, II = a i a i κ j+i κ j+i x j+i x j+i j=0 i i } {{} off diagonal...summation in the index is modulo d... ) P a,κ ( M a,k D κ x 2 2 (1+ε)k P a (I (1+ε/2)k)+P a,κ (II εk/2)

23 Estimates of I: P a (I (1 + ε/2)k) Lemma of B. Laurent and P. Massart...or any other variant of Bernstein s inequality Exponential concentration of Z = D α i (ai 2 1), i=1 where a i are i.i.d. normal variables and α i are nonnegative real numbers. Then for any t > 0 P(Z 2 α 2 t + 2 α t) exp( t), P(Z 2 α 2 t) exp( t). α i := k 1 j=0 x2 j+i, α 1 = k, α 1 and α 2 k.

24 Estimates of II: Decoupling lemma of Bourgain and Tzafriri: Let ξ 0,...,ξ d 1 be independent random variables with E ξ 0 = = E ξ d 1 = 0 and let {x i,j } d 1 i,j=0 be a double sequence of real numbers. Then for 1 p < E p x i,j ξ i ξ j 4 p E x i,j ξ i ξ j p, i j where (ξ 0,...,ξ d 1 ) denotes an independent copy of (ξ 0,...,ξ d 1 ). i j

25 Further tools: Two times Khintchine s inequalities and ( E a,a k 1 a j a j p) 1/p p(k + p), j=0 for both a Bernoulli or Gaussian variables.

26 The role of D κ k d, a 0,...,a d 1 independent normal variables 2-stability: x = 1 d (1,...,1), b := d 1 j=0 P a ( M a,k x 2 2 > (1 + ε)k depends neither on k nor on d (d 1 M a,k x 2 2 = k a ) j 2 d a j d N(0,1) j=0 ) ( ) = P b b 2 > (1 + ε)

27 log 2 n Let x 1,...,x n be arbitrary points in R d, ε (0, 1 2 ), k = O(ε 2 log 2 n), a = (a 0,...,a d 1 ) be independent normally distributed variables, M a,k and D κ be as above and f (x) = 1 k M a,k D κ x. Then with probability at least 2/3 the following holds (1 ε) x i x j 2 2 f (x i ) f (x j ) 2 2 (1+ε) x i x j 2 2, i,j = 1,...,n.

28 Fourier methods F - unitary discrete Fourier transform, F : C d C d Every circulant matrix may be diagonalised by F and F 1 M a,d x = Fdiag( dfa)f 1 x. The singular values are the square roots of the eigenvalues of M a,d M a,d = Fdiag( dfa)diag( dfa)f 1 = Fdiag(d Fa 2 )F 1 i.e. d Fa.

29 Strategy of the proof of log 2 n Concentration inequalities for all x = x i x j x i x j 2 ( P a Ma,k D κ x 2 2 2(1 + ε)k ) exp ( ckε2 ), log n ( P a Ma,k D κ x 2 2 2(1 ε)k ) exp ( ckε2 ). log n From this, the result follows again by a union bound.

30 Let x 2 = 1. y j := S j (D κ x) C d, j = 0,...,k 1, where S is the shift operator S : C d C d, S(z 0,...,z d 1 ) = (z 1,...,z d 1,z 0 ). Y...k d matrix with rows y 0,...,y k 1. Note, that M a,k D κ x 2 2 = Ya 2 2 Hence, P ( M a,k D κ x 2 2 2(1 + ε)k) = P ( Ya 2 2 2(1 + ε)k).

31 Let Y = UΣV be the singular value decomposition of Y. Then Ya 2 2 = UΣVa 2 2 = ΣVa 2 2 = Σb 2 2, where b := Va is a k-dimensional vector of independent normal variables. Hence, P ( M a,k D κ x 2 2 (1 + ε)k ) = P (k 1 λ 2 j bj 2 (1 + ε)k ), where λ j are the singular values of Y. Lemma of B. Laurent and P. Massart: Estimate λ 4 and λ! j=0 λ 2 2 = Y F = k and λ 2 c log n implies λ 4 4 c k log n.

32 Related results: N. Ailon and E. Liberty, Almost optimal unrestricted fast Johnson-Lindenstrauss transform, Random partial Fourier transform, k = O(ε 4 log n log 4 d) H. Rauhut, J. Romberg and J. Tropp, Restricted isometries for partial random circulant matrices, Random circulant matrices { } k max ε 1 log 3/2 n log 3/2 d,ε 2 log n log 4 d.

33 Approximate nearest neighbors The nearest neighbor problem: Given P = {x 1,...,x n } in a metric space X, preprocess P so as to efficiently find the minimizer of min d(x i,q), q X. i=1,...,n Donald Knuth in vol. 3 of The Art of Computer Programming called it the post-office problem Naive algorithm: compare all the distances - no preprocessing. The approximate nearest neighbor problem: Given P = {x 1,...,x n } in a metric space X and ε > 0, preprocess P so as to efficiently find p P, such that d(p,q) (1 + ε)d(p,q), p X.

34 P = {x 1,...,x n } X = R d, hashing functions: Choose randomly v R d and pre-compute h v (i) := v,x i, i = 1,...,n Find argmin i=1,...,n v,x i q. Iterate over different v 1,v 2,...

35 Dynamical systems J-L-transforms are linear and (almost) preserve distances. We consider non-linear dynamical systems, where the non-linearity depends only on distances of the agents: ẋ i (t) = f i (Dx(t)) + N f ij (Dx(t))x j (t) j=1 N N - number of agents, x(t) = (x 1 (t),...,x N (t)) R d N, where x i : [0,T] R d,i = 1,...,N, f i : R N N R d,i = 1,...,N, f ij : R N N R,i,j = 1,...,N, D : R d N R N N, Dx := ( x i x j l d 2 ),i,j = 1,...,N is the adjacency matrix.

36 Euler s method: x i (0) = x 0 (i),i = 1,...,N and N := xi n + h f i (Dx n ) + f ij (Dx n )xj n,n = 0,...,n 0 1. x n+1 i j=1 We project the system using J-L-transform M R k d : y i (0) = Mx 0 (i),i = 1,...,N and N := yi n +h Mf i (D y n ) + y n+1 i j=1 Mf ij (D y n )y n j If f i and f i,j are Lipschitz: error estimates for y n Mx n.,n = 0,...,n 0 1.

37 References: W. B. Johnson and J. Lindenstrauss, Extensions of Lipschitz mappings into a Hilbert space. Contem. Math., 26: , 1984 S. Dasgupta and A. Gupta, An elementary proof of a theorem of Johnson and Lindenstrauss. Random Struct. & Algorithms, 22:60-65, N. Ailon and B. Chazelle, Approximate nearest neighbors and the fast Johnson-Lindenstrauss transform. In Proc. 38th Annual ACM Symposium on Theory of Computing, A. Hinrichs and J. Vybíral, Johnson-Lindenstrauss lemma for circulant matrices, to appear in Random Struct. & Algorithms J. Vybíral, A variant of the Johnson-Lindenstrauss lemma for circulant matrices, to appear in J. Funct. Anal.

Some Useful Background for Talk on the Fast Johnson-Lindenstrauss Transform

Some Useful Background for Talk on the Fast Johnson-Lindenstrauss Transform Some Useful Background for Talk on the Fast Johnson-Lindenstrauss Transform Nir Ailon May 22, 2007 This writeup includes very basic background material for the talk on the Fast Johnson Lindenstrauss Transform

More information

Randomized Algorithms

Randomized Algorithms Randomized Algorithms Saniv Kumar, Google Research, NY EECS-6898, Columbia University - Fall, 010 Saniv Kumar 9/13/010 EECS6898 Large Scale Machine Learning 1 Curse of Dimensionality Gaussian Mixture Models

More information

Non-Asymptotic Theory of Random Matrices Lecture 4: Dimension Reduction Date: January 16, 2007

Non-Asymptotic Theory of Random Matrices Lecture 4: Dimension Reduction Date: January 16, 2007 Non-Asymptotic Theory of Random Matrices Lecture 4: Dimension Reduction Date: January 16, 2007 Lecturer: Roman Vershynin Scribe: Matthew Herman 1 Introduction Consider the set X = {n points in R N } where

More information

Faster Johnson-Lindenstrauss style reductions

Faster Johnson-Lindenstrauss style reductions Faster Johnson-Lindenstrauss style reductions Aditya Menon August 23, 2007 Outline 1 Introduction Dimensionality reduction The Johnson-Lindenstrauss Lemma Speeding up computation 2 The Fast Johnson-Lindenstrauss

More information

Fast Random Projections

Fast Random Projections Fast Random Projections Edo Liberty 1 September 18, 2007 1 Yale University, New Haven CT, supported by AFOSR and NGA (www.edoliberty.com) Advised by Steven Zucker. About This talk will survey a few random

More information

Methods for sparse analysis of high-dimensional data, II

Methods for sparse analysis of high-dimensional data, II Methods for sparse analysis of high-dimensional data, II Rachel Ward May 23, 2011 High dimensional data with low-dimensional structure 300 by 300 pixel images = 90, 000 dimensions 2 / 47 High dimensional

More information

Fast Dimension Reduction

Fast Dimension Reduction Fast Dimension Reduction Nir Ailon 1 Edo Liberty 2 1 Google Research 2 Yale University Introduction Lemma (Johnson, Lindenstrauss (1984)) A random projection Ψ preserves all ( n 2) distances up to distortion

More information

Sparse Johnson-Lindenstrauss Transforms

Sparse Johnson-Lindenstrauss Transforms Sparse Johnson-Lindenstrauss Transforms Jelani Nelson MIT May 24, 211 joint work with Daniel Kane (Harvard) Metric Johnson-Lindenstrauss lemma Metric JL (MJL) Lemma, 1984 Every set of n points in Euclidean

More information

MAT 585: Johnson-Lindenstrauss, Group testing, and Compressed Sensing

MAT 585: Johnson-Lindenstrauss, Group testing, and Compressed Sensing MAT 585: Johnson-Lindenstrauss, Group testing, and Compressed Sensing Afonso S. Bandeira April 9, 2015 1 The Johnson-Lindenstrauss Lemma Suppose one has n points, X = {x 1,..., x n }, in R d with d very

More information

Methods for sparse analysis of high-dimensional data, II

Methods for sparse analysis of high-dimensional data, II Methods for sparse analysis of high-dimensional data, II Rachel Ward May 26, 2011 High dimensional data with low-dimensional structure 300 by 300 pixel images = 90, 000 dimensions 2 / 55 High dimensional

More information

Random Methods for Linear Algebra

Random Methods for Linear Algebra Gittens gittens@acm.caltech.edu Applied and Computational Mathematics California Institue of Technology October 2, 2009 Outline The Johnson-Lindenstrauss Transform 1 The Johnson-Lindenstrauss Transform

More information

Fast Dimension Reduction

Fast Dimension Reduction Fast Dimension Reduction MMDS 2008 Nir Ailon Google Research NY Fast Dimension Reduction Using Rademacher Series on Dual BCH Codes (with Edo Liberty) The Fast Johnson Lindenstrauss Transform (with Bernard

More information

Random projections. 1 Introduction. 2 Dimensionality reduction. Lecture notes 5 February 29, 2016

Random projections. 1 Introduction. 2 Dimensionality reduction. Lecture notes 5 February 29, 2016 Lecture notes 5 February 9, 016 1 Introduction Random projections Random projections are a useful tool in the analysis and processing of high-dimensional data. We will analyze two applications that use

More information

Dimensionality Reduction Notes 3

Dimensionality Reduction Notes 3 Dimensionality Reduction Notes 3 Jelani Nelson minilek@seas.harvard.edu August 13, 2015 1 Gordon s theorem Let T be a finite subset of some normed vector space with norm X. We say that a sequence T 0 T

More information

18.409: Topics in TCS: Embeddings of Finite Metric Spaces. Lecture 6

18.409: Topics in TCS: Embeddings of Finite Metric Spaces. Lecture 6 Massachusetts Institute of Technology Lecturer: Michel X. Goemans 8.409: Topics in TCS: Embeddings of Finite Metric Spaces September 7, 006 Scribe: Benjamin Rossman Lecture 6 Today we loo at dimension

More information

Sparser Johnson-Lindenstrauss Transforms

Sparser Johnson-Lindenstrauss Transforms Sparser Johnson-Lindenstrauss Transforms Jelani Nelson Princeton February 16, 212 joint work with Daniel Kane (Stanford) Random Projections x R d, d huge store y = Sx, where S is a k d matrix (compression)

More information

Lecture 3. Random Fourier measurements

Lecture 3. Random Fourier measurements Lecture 3. Random Fourier measurements 1 Sampling from Fourier matrices 2 Law of Large Numbers and its operator-valued versions 3 Frames. Rudelson s Selection Theorem Sampling from Fourier matrices Our

More information

16 Embeddings of the Euclidean metric

16 Embeddings of the Euclidean metric 16 Embeddings of the Euclidean metric In today s lecture, we will consider how well we can embed n points in the Euclidean metric (l 2 ) into other l p metrics. More formally, we ask the following question.

More information

Acceleration of Randomized Kaczmarz Method

Acceleration of Randomized Kaczmarz Method Acceleration of Randomized Kaczmarz Method Deanna Needell [Joint work with Y. Eldar] Stanford University BIRS Banff, March 2011 Problem Background Setup Setup Let Ax = b be an overdetermined consistent

More information

Randomized Algorithms

Randomized Algorithms Randomized Algorithms 南京大学 尹一通 Martingales Definition: A sequence of random variables X 0, X 1,... is a martingale if for all i > 0, E[X i X 0,...,X i1 ] = X i1 x 0, x 1,...,x i1, E[X i X 0 = x 0, X 1

More information

A fast randomized algorithm for overdetermined linear least-squares regression

A fast randomized algorithm for overdetermined linear least-squares regression A fast randomized algorithm for overdetermined linear least-squares regression Vladimir Rokhlin and Mark Tygert Technical Report YALEU/DCS/TR-1403 April 28, 2008 Abstract We introduce a randomized algorithm

More information

On positive positive-definite functions and Bochner s Theorem

On positive positive-definite functions and Bochner s Theorem On positive positive-definite functions and Bochner s Theorem Aicke Hinrichs Mathematisches Institut, Universität Jena Ernst-Abbe-Platz, 07740 Jena, Germany email: a.hinrichs@uni-jena.de Jan Vybíral RICAM,

More information

Yale university technical report #1402.

Yale university technical report #1402. The Mailman algorithm: a note on matrix vector multiplication Yale university technical report #1402. Edo Liberty Computer Science Yale University New Haven, CT Steven W. Zucker Computer Science and Appled

More information

Accelerated Dense Random Projections

Accelerated Dense Random Projections 1 Advisor: Steven Zucker 1 Yale University, Department of Computer Science. Dimensionality reduction (1 ε) xi x j 2 Ψ(xi ) Ψ(x j ) 2 (1 + ε) xi x j 2 ( n 2) distances are ε preserved Target dimension k

More information

CS on CS: Computer Science insights into Compresive Sensing (and vice versa) Piotr Indyk MIT

CS on CS: Computer Science insights into Compresive Sensing (and vice versa) Piotr Indyk MIT CS on CS: Computer Science insights into Compresive Sensing (and vice versa) Piotr Indyk MIT Sparse Approximations Goal: approximate a highdimensional vector x by x that is sparse, i.e., has few nonzero

More information

Multivariate Statistics Random Projections and Johnson-Lindenstrauss Lemma

Multivariate Statistics Random Projections and Johnson-Lindenstrauss Lemma Multivariate Statistics Random Projections and Johnson-Lindenstrauss Lemma Suppose again we have n sample points x,..., x n R p. The data-point x i R p can be thought of as the i-th row X i of an n p-dimensional

More information

Supremum of simple stochastic processes

Supremum of simple stochastic processes Subspace embeddings Daniel Hsu COMS 4772 1 Supremum of simple stochastic processes 2 Recap: JL lemma JL lemma. For any ε (0, 1/2), point set S R d of cardinality 16 ln n S = n, and k N such that k, there

More information

JOHNSON-LINDENSTRAUSS TRANSFORMATION AND RANDOM PROJECTION

JOHNSON-LINDENSTRAUSS TRANSFORMATION AND RANDOM PROJECTION JOHNSON-LINDENSTRAUSS TRANSFORMATION AND RANDOM PROJECTION LONG CHEN ABSTRACT. We give a brief survey of Johnson-Lindenstrauss lemma. CONTENTS 1. Introduction 1 2. JL Transform 4 2.1. An Elementary Proof

More information

Optimal compression of approximate Euclidean distances

Optimal compression of approximate Euclidean distances Optimal compression of approximate Euclidean distances Noga Alon 1 Bo az Klartag 2 Abstract Let X be a set of n points of norm at most 1 in the Euclidean space R k, and suppose ε > 0. An ε-distance sketch

More information

Limitations in Approximating RIP

Limitations in Approximating RIP Alok Puranik Mentor: Adrian Vladu Fifth Annual PRIMES Conference, 2015 Outline 1 Background The Problem Motivation Construction Certification 2 Planted model Planting eigenvalues Analysis Distinguishing

More information

Kernel Method: Data Analysis with Positive Definite Kernels

Kernel Method: Data Analysis with Positive Definite Kernels Kernel Method: Data Analysis with Positive Definite Kernels 2. Positive Definite Kernel and Reproducing Kernel Hilbert Space Kenji Fukumizu The Institute of Statistical Mathematics. Graduate University

More information

CS 229r: Algorithms for Big Data Fall Lecture 17 10/28

CS 229r: Algorithms for Big Data Fall Lecture 17 10/28 CS 229r: Algorithms for Big Data Fall 2015 Prof. Jelani Nelson Lecture 17 10/28 Scribe: Morris Yau 1 Overview In the last lecture we defined subspace embeddings a subspace embedding is a linear transformation

More information

Using the Johnson-Lindenstrauss lemma in linear and integer programming

Using the Johnson-Lindenstrauss lemma in linear and integer programming Using the Johnson-Lindenstrauss lemma in linear and integer programming Vu Khac Ky 1, Pierre-Louis Poirion, Leo Liberti LIX, École Polytechnique, F-91128 Palaiseau, France Email:{vu,poirion,liberti}@lix.polytechnique.fr

More information

Advances in Manifold Learning Presented by: Naku Nak l Verm r a June 10, 2008

Advances in Manifold Learning Presented by: Naku Nak l Verm r a June 10, 2008 Advances in Manifold Learning Presented by: Nakul Verma June 10, 008 Outline Motivation Manifolds Manifold Learning Random projection of manifolds for dimension reduction Introduction to random projections

More information

Very Sparse Random Projections

Very Sparse Random Projections Very Sparse Random Projections Ping Li, Trevor Hastie and Kenneth Church [KDD 06] Presented by: Aditya Menon UCSD March 4, 2009 Presented by: Aditya Menon (UCSD) Very Sparse Random Projections March 4,

More information

Lecture 13 October 6, Covering Numbers and Maurey s Empirical Method

Lecture 13 October 6, Covering Numbers and Maurey s Empirical Method CS 395T: Sublinear Algorithms Fall 2016 Prof. Eric Price Lecture 13 October 6, 2016 Scribe: Kiyeon Jeon and Loc Hoang 1 Overview In the last lecture we covered the lower bound for p th moment (p > 2) and

More information

Compressed Sensing: Lecture I. Ronald DeVore

Compressed Sensing: Lecture I. Ronald DeVore Compressed Sensing: Lecture I Ronald DeVore Motivation Compressed Sensing is a new paradigm for signal/image/function acquisition Motivation Compressed Sensing is a new paradigm for signal/image/function

More information

Dense Fast Random Projections and Lean Walsh Transforms

Dense Fast Random Projections and Lean Walsh Transforms Dense Fast Random Projections and Lean Walsh Transforms Edo Liberty, Nir Ailon, and Amit Singer Abstract. Random projection methods give distributions over k d matrices such that if a matrix Ψ (chosen

More information

Fast Random Projections using Lean Walsh Transforms Yale University Technical report #1390

Fast Random Projections using Lean Walsh Transforms Yale University Technical report #1390 Fast Random Projections using Lean Walsh Transforms Yale University Technical report #1390 Edo Liberty Nir Ailon Amit Singer Abstract We present a k d random projection matrix that is applicable to vectors

More information

A Simple Proof of the Restricted Isometry Property for Random Matrices

A Simple Proof of the Restricted Isometry Property for Random Matrices DOI 10.1007/s00365-007-9003-x A Simple Proof of the Restricted Isometry Property for Random Matrices Richard Baraniuk Mark Davenport Ronald DeVore Michael Wakin Received: 17 May 006 / Revised: 18 January

More information

Guaranteed Minimum-Rank Solutions of Linear Matrix Equations via Nuclear Norm Minimization

Guaranteed Minimum-Rank Solutions of Linear Matrix Equations via Nuclear Norm Minimization Forty-Fifth Annual Allerton Conference Allerton House, UIUC, Illinois, USA September 26-28, 27 WeA3.2 Guaranteed Minimum-Rank Solutions of Linear Matrix Equations via Nuclear Norm Minimization Benjamin

More information

Reconstruction from Anisotropic Random Measurements

Reconstruction from Anisotropic Random Measurements Reconstruction from Anisotropic Random Measurements Mark Rudelson and Shuheng Zhou The University of Michigan, Ann Arbor Coding, Complexity, and Sparsity Workshop, 013 Ann Arbor, Michigan August 7, 013

More information

Improved Bounds on the Dot Product under Random Projection and Random Sign Projection

Improved Bounds on the Dot Product under Random Projection and Random Sign Projection Improved Bounds on the Dot Product under Random Projection and Random Sign Projection Ata Kabán School of Computer Science The University of Birmingham Birmingham B15 2TT, UK http://www.cs.bham.ac.uk/

More information

Strengthened Sobolev inequalities for a random subspace of functions

Strengthened Sobolev inequalities for a random subspace of functions Strengthened Sobolev inequalities for a random subspace of functions Rachel Ward University of Texas at Austin April 2013 2 Discrete Sobolev inequalities Proposition (Sobolev inequality for discrete images)

More information

Near Ideal Behavior of a Modified Elastic Net Algorithm in Compressed Sensing

Near Ideal Behavior of a Modified Elastic Net Algorithm in Compressed Sensing Near Ideal Behavior of a Modified Elastic Net Algorithm in Compressed Sensing M. Vidyasagar Cecil & Ida Green Chair The University of Texas at Dallas M.Vidyasagar@utdallas.edu www.utdallas.edu/ m.vidyasagar

More information

Lecture 18: March 15

Lecture 18: March 15 CS71 Randomness & Computation Spring 018 Instructor: Alistair Sinclair Lecture 18: March 15 Disclaimer: These notes have not been subjected to the usual scrutiny accorded to formal publications. They may

More information

1 Dimension Reduction in Euclidean Space

1 Dimension Reduction in Euclidean Space CSIS0351/8601: Randomized Algorithms Lecture 6: Johnson-Lindenstrauss Lemma: Dimension Reduction Lecturer: Hubert Chan Date: 10 Oct 011 hese lecture notes are supplementary materials for the lectures.

More information

The Johnson-Lindenstrauss Lemma

The Johnson-Lindenstrauss Lemma The Johnson-Lindenstrauss Lemma Kevin (Min Seong) Park MAT477 Introduction The Johnson-Lindenstrauss Lemma was first introduced in the paper Extensions of Lipschitz mappings into a Hilbert Space by William

More information

Exponential decay of reconstruction error from binary measurements of sparse signals

Exponential decay of reconstruction error from binary measurements of sparse signals Exponential decay of reconstruction error from binary measurements of sparse signals Deanna Needell Joint work with R. Baraniuk, S. Foucart, Y. Plan, and M. Wootters Outline Introduction Mathematical Formulation

More information

A fast randomized algorithm for approximating an SVD of a matrix

A fast randomized algorithm for approximating an SVD of a matrix A fast randomized algorithm for approximating an SVD of a matrix Joint work with Franco Woolfe, Edo Liberty, and Vladimir Rokhlin Mark Tygert Program in Applied Mathematics Yale University Place July 17,

More information

sample lectures: Compressed Sensing, Random Sampling I & II

sample lectures: Compressed Sensing, Random Sampling I & II P. Jung, CommIT, TU-Berlin]: sample lectures Compressed Sensing / Random Sampling I & II 1/68 sample lectures: Compressed Sensing, Random Sampling I & II Peter Jung Communications and Information Theory

More information

CS168: The Modern Algorithmic Toolbox Lecture #4: Dimensionality Reduction

CS168: The Modern Algorithmic Toolbox Lecture #4: Dimensionality Reduction CS168: The Modern Algorithmic Toolbox Lecture #4: Dimensionality Reduction Tim Roughgarden & Gregory Valiant April 12, 2017 1 The Curse of Dimensionality in the Nearest Neighbor Problem Lectures #1 and

More information

Johnson-Lindenstrauss, Concentration and applications to Support Vector Machines and Kernels

Johnson-Lindenstrauss, Concentration and applications to Support Vector Machines and Kernels Johnson-Lindenstrauss, Concentration and applications to Support Vector Machines and Kernels Devdatt Dubhashi Department of Computer Science and Engineering, Chalmers University, dubhashi@chalmers.se Functional

More information

arxiv: v1 [math.pr] 22 May 2008

arxiv: v1 [math.pr] 22 May 2008 THE LEAST SINGULAR VALUE OF A RANDOM SQUARE MATRIX IS O(n 1/2 ) arxiv:0805.3407v1 [math.pr] 22 May 2008 MARK RUDELSON AND ROMAN VERSHYNIN Abstract. Let A be a matrix whose entries are real i.i.d. centered

More information

Sketching as a Tool for Numerical Linear Algebra

Sketching as a Tool for Numerical Linear Algebra Sketching as a Tool for Numerical Linear Algebra David P. Woodruff presented by Sepehr Assadi o(n) Big Data Reading Group University of Pennsylvania February, 2015 Sepehr Assadi (Penn) Sketching for Numerical

More information

Nearest Neighbor Preserving Embeddings

Nearest Neighbor Preserving Embeddings Nearest Neighbor Preserving Embeddings Piotr Indyk MIT Assaf Naor Microsoft Research Abstract In this paper we introduce the notion of nearest neighbor preserving embeddings. These are randomized embeddings

More information

STAT 200C: High-dimensional Statistics

STAT 200C: High-dimensional Statistics STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 59 Classical case: n d. Asymptotic assumption: d is fixed and n. Basic tools: LLN and CLT. High-dimensional setting: n d, e.g. n/d

More information

AN INTRODUCTION TO COMPRESSIVE SENSING

AN INTRODUCTION TO COMPRESSIVE SENSING AN INTRODUCTION TO COMPRESSIVE SENSING Rodrigo B. Platte School of Mathematical and Statistical Sciences APM/EEE598 Reverse Engineering of Complex Dynamical Networks OUTLINE 1 INTRODUCTION 2 INCOHERENCE

More information

Constructing Explicit RIP Matrices and the Square-Root Bottleneck

Constructing Explicit RIP Matrices and the Square-Root Bottleneck Constructing Explicit RIP Matrices and the Square-Root Bottleneck Ryan Cinoman July 18, 2018 Ryan Cinoman Constructing Explicit RIP Matrices July 18, 2018 1 / 36 Outline 1 Introduction 2 Restricted Isometry

More information

Invertibility of symmetric random matrices

Invertibility of symmetric random matrices Invertibility of symmetric random matrices Roman Vershynin University of Michigan romanv@umich.edu February 1, 2011; last revised March 16, 2012 Abstract We study n n symmetric random matrices H, possibly

More information

Elements of Positive Definite Kernel and Reproducing Kernel Hilbert Space

Elements of Positive Definite Kernel and Reproducing Kernel Hilbert Space Elements of Positive Definite Kernel and Reproducing Kernel Hilbert Space Statistical Inference with Reproducing Kernel Hilbert Space Kenji Fukumizu Institute of Statistical Mathematics, ROIS Department

More information

Random regular digraphs: singularity and spectrum

Random regular digraphs: singularity and spectrum Random regular digraphs: singularity and spectrum Nick Cook, UCLA Probability Seminar, Stanford University November 2, 2015 Universality Circular law Singularity probability Talk outline 1 Universality

More information

Compressibility of Infinite Sequences and its Interplay with Compressed Sensing Recovery

Compressibility of Infinite Sequences and its Interplay with Compressed Sensing Recovery Compressibility of Infinite Sequences and its Interplay with Compressed Sensing Recovery Jorge F. Silva and Eduardo Pavez Department of Electrical Engineering Information and Decision Systems Group Universidad

More information

Sketching as a Tool for Numerical Linear Algebra All Lectures. David Woodruff IBM Almaden

Sketching as a Tool for Numerical Linear Algebra All Lectures. David Woodruff IBM Almaden Sketching as a Tool for Numerical Linear Algebra All Lectures David Woodruff IBM Almaden Massive data sets Examples Internet traffic logs Financial data etc. Algorithms Want nearly linear time or less

More information

Review of similarity transformation and Singular Value Decomposition

Review of similarity transformation and Singular Value Decomposition Review of similarity transformation and Singular Value Decomposition Nasser M Abbasi Applied Mathematics Department, California State University, Fullerton July 8 7 page compiled on June 9, 5 at 9:5pm

More information

Invertibility of random matrices

Invertibility of random matrices University of Michigan February 2011, Princeton University Origins of Random Matrix Theory Statistics (Wishart matrices) PCA of a multivariate Gaussian distribution. [Gaël Varoquaux s blog gael-varoquaux.info]

More information

Randomness Efficient Fast-Johnson-Lindenstrauss Transform with Applications in Differential Privacy

Randomness Efficient Fast-Johnson-Lindenstrauss Transform with Applications in Differential Privacy Randomness Efficient Fast-Johnson-Lindenstrauss Transform with Applications in Differential Privacy Jalaj Upadhyay Cheriton School of Computer Science University of Waterloo jkupadhy@csuwaterlooca arxiv:14102470v6

More information

The Singular Value Decomposition

The Singular Value Decomposition The Singular Value Decomposition Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) SVD Fall 2015 1 / 13 Review of Key Concepts We review some key definitions and results about matrices that will

More information

Model Counting for Logical Theories

Model Counting for Logical Theories Model Counting for Logical Theories Wednesday Dmitry Chistikov Rayna Dimitrova Department of Computer Science University of Oxford, UK Max Planck Institute for Software Systems (MPI-SWS) Kaiserslautern

More information

Linear Algebra Review

Linear Algebra Review January 29, 2013 Table of contents Metrics Metric Given a space X, then d : X X R + 0 and z in X if: d(x, y) = 0 is equivalent to x = y d(x, y) = d(y, x) d(x, y) d(x, z) + d(z, y) is a metric is for all

More information

Homework 1. Yuan Yao. September 18, 2011

Homework 1. Yuan Yao. September 18, 2011 Homework 1 Yuan Yao September 18, 2011 1. Singular Value Decomposition: The goal of this exercise is to refresh your memory about the singular value decomposition and matrix norms. A good reference to

More information

Lecture Notes 1: Vector spaces

Lecture Notes 1: Vector spaces Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector

More information

Randomized Numerical Linear Algebra: Review and Progresses

Randomized Numerical Linear Algebra: Review and Progresses ized ized SVD ized : Review and Progresses Zhihua Department of Computer Science and Engineering Shanghai Jiao Tong University The 12th China Workshop on Machine Learning and Applications Xi an, November

More information

Introduction to Compressed Sensing

Introduction to Compressed Sensing Introduction to Compressed Sensing Alejandro Parada, Gonzalo Arce University of Delaware August 25, 2016 Motivation: Classical Sampling 1 Motivation: Classical Sampling Issues Some applications Radar Spectral

More information

Homework 2 Foundations of Computational Math 2 Spring 2019

Homework 2 Foundations of Computational Math 2 Spring 2019 Homework 2 Foundations of Computational Math 2 Spring 2019 Problem 2.1 (2.1.a) Suppose (v 1,λ 1 )and(v 2,λ 2 ) are eigenpairs for a matrix A C n n. Show that if λ 1 λ 2 then v 1 and v 2 are linearly independent.

More information

Lecture Notes 9: Constrained Optimization

Lecture Notes 9: Constrained Optimization Optimization-based data analysis Fall 017 Lecture Notes 9: Constrained Optimization 1 Compressed sensing 1.1 Underdetermined linear inverse problems Linear inverse problems model measurements of the form

More information

Model-Based Compressive Sensing for Signal Ensembles. Marco F. Duarte Volkan Cevher Richard G. Baraniuk

Model-Based Compressive Sensing for Signal Ensembles. Marco F. Duarte Volkan Cevher Richard G. Baraniuk Model-Based Compressive Sensing for Signal Ensembles Marco F. Duarte Volkan Cevher Richard G. Baraniuk Concise Signal Structure Sparse signal: only K out of N coordinates nonzero model: union of K-dimensional

More information

A Power Efficient Sensing/Communication Scheme: Joint Source-Channel-Network Coding by Using Compressive Sensing

A Power Efficient Sensing/Communication Scheme: Joint Source-Channel-Network Coding by Using Compressive Sensing Forty-Ninth Annual Allerton Conference Allerton House, UIUC, Illinois, USA September 28-30, 20 A Power Efficient Sensing/Communication Scheme: Joint Source-Channel-Network Coding by Using Compressive Sensing

More information

Concentration of Measure Inequalities for. Toeplitz Matrices with Applications

Concentration of Measure Inequalities for. Toeplitz Matrices with Applications Concentration of Measure Inequalities for 1 Toeplitz Matrices with Applications Borhan M. Sanandaji, Tyrone L. Vincent, and Michael B. Wakin Abstract We derive Concentration of Measure (CoM) inequalities

More information

Notes taken by Costis Georgiou revised by Hamed Hatami

Notes taken by Costis Georgiou revised by Hamed Hatami CSC414 - Metric Embeddings Lecture 6: Reductions that preserve volumes and distance to affine spaces & Lower bound techniques for distortion when embedding into l Notes taken by Costis Georgiou revised

More information

STA141C: Big Data & High Performance Statistical Computing

STA141C: Big Data & High Performance Statistical Computing STA141C: Big Data & High Performance Statistical Computing Numerical Linear Algebra Background Cho-Jui Hsieh UC Davis May 15, 2018 Linear Algebra Background Vectors A vector has a direction and a magnitude

More information

Sparse Legendre expansions via l 1 minimization

Sparse Legendre expansions via l 1 minimization Sparse Legendre expansions via l 1 minimization Rachel Ward, Courant Institute, NYU Joint work with Holger Rauhut, Hausdorff Center for Mathematics, Bonn, Germany. June 8, 2010 Outline Sparse recovery

More information

Approximate Spectral Clustering via Randomized Sketching

Approximate Spectral Clustering via Randomized Sketching Approximate Spectral Clustering via Randomized Sketching Christos Boutsidis Yahoo! Labs, New York Joint work with Alex Gittens (Ebay), Anju Kambadur (IBM) The big picture: sketch and solve Tradeoff: Speed

More information

Lecture 15: Random Projections

Lecture 15: Random Projections Lecture 15: Random Projections Introduction to Learning and Analysis of Big Data Kontorovich and Sabato (BGU) Lecture 15 1 / 11 Review of PCA Unsupervised learning technique Performs dimensionality reduction

More information

A Unified Theorem on SDP Rank Reduction. yyye

A Unified Theorem on SDP Rank Reduction.   yyye SDP Rank Reduction Yinyu Ye, EURO XXII 1 A Unified Theorem on SDP Rank Reduction Yinyu Ye Department of Management Science and Engineering and Institute of Computational and Mathematical Engineering Stanford

More information

The Johnson-Lindenstrauss Lemma in Linear Programming

The Johnson-Lindenstrauss Lemma in Linear Programming The Johnson-Lindenstrauss Lemma in Linear Programming Leo Liberti, Vu Khac Ky, Pierre-Louis Poirion CNRS LIX Ecole Polytechnique, France Aussois COW 2016 The gist Goal: solving very large LPs min{c x Ax

More information

arxiv: v1 [cs.ds] 30 May 2010

arxiv: v1 [cs.ds] 30 May 2010 ALMOST OPTIMAL UNRESTRICTED FAST JOHNSON-LINDENSTRAUSS TRANSFORM arxiv:005.553v [cs.ds] 30 May 200 NIR AILON AND EDO LIBERTY Abstract. The problems of random projections and sparse reconstruction have

More information

Matrix Theory. A.Holst, V.Ufnarovski

Matrix Theory. A.Holst, V.Ufnarovski Matrix Theory AHolst, VUfnarovski 55 HINTS AND ANSWERS 9 55 Hints and answers There are two different approaches In the first one write A as a block of rows and note that in B = E ij A all rows different

More information

CS 229r: Algorithms for Big Data Fall Lecture 19 Nov 5

CS 229r: Algorithms for Big Data Fall Lecture 19 Nov 5 CS 229r: Algorithms for Big Data Fall 215 Prof. Jelani Nelson Lecture 19 Nov 5 Scribe: Abdul Wasay 1 Overview In the last lecture, we started discussing the problem of compressed sensing where we are given

More information

Quantum Computing Lecture 2. Review of Linear Algebra

Quantum Computing Lecture 2. Review of Linear Algebra Quantum Computing Lecture 2 Review of Linear Algebra Maris Ozols Linear algebra States of a quantum system form a vector space and their transformations are described by linear operators Vector spaces

More information

Compressive Sensing and Beyond

Compressive Sensing and Beyond Compressive Sensing and Beyond Sohail Bahmani Gerorgia Tech. Signal Processing Compressed Sensing Signal Models Classics: bandlimited The Sampling Theorem Any signal with bandwidth B can be recovered

More information

Fast Parallel Estimation of High Dimensional Information Theoretical Quantities with Low Dimensional Random Projection Ensembles

Fast Parallel Estimation of High Dimensional Information Theoretical Quantities with Low Dimensional Random Projection Ensembles Fast Parallel Estimation of High Dimensional Information Theoretical Quantities with Low Dimensional Random Projection Ensembles Zoltán Szabó and András Lőrincz Department of Information Systems, Eötvös

More information

Constrained optimization

Constrained optimization Constrained optimization DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_fall17/index.html Carlos Fernandez-Granda Compressed sensing Convex constrained

More information

THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR

THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR WEN LI AND MICHAEL K. NG Abstract. In this paper, we study the perturbation bound for the spectral radius of an m th - order n-dimensional

More information

MANY scientific computations, signal processing, data analysis and machine learning applications lead to large dimensional

MANY scientific computations, signal processing, data analysis and machine learning applications lead to large dimensional Low rank approximation and decomposition of large matrices using error correcting codes Shashanka Ubaru, Arya Mazumdar, and Yousef Saad 1 arxiv:1512.09156v3 [cs.it] 15 Jun 2017 Abstract Low rank approximation

More information

High Dimensional Geometry, Curse of Dimensionality, Dimension Reduction

High Dimensional Geometry, Curse of Dimensionality, Dimension Reduction Chapter 11 High Dimensional Geometry, Curse of Dimensionality, Dimension Reduction High-dimensional vectors are ubiquitous in applications (gene expression data, set of movies watched by Netflix customer,

More information

Sketching as a Tool for Numerical Linear Algebra

Sketching as a Tool for Numerical Linear Algebra Sketching as a Tool for Numerical Linear Algebra (Part 2) David P. Woodruff presented by Sepehr Assadi o(n) Big Data Reading Group University of Pennsylvania February, 2015 Sepehr Assadi (Penn) Sketching

More information

Sparse Recovery with Pre-Gaussian Random Matrices

Sparse Recovery with Pre-Gaussian Random Matrices Sparse Recovery with Pre-Gaussian Random Matrices Simon Foucart Laboratoire Jacques-Louis Lions Université Pierre et Marie Curie Paris, 75013, France Ming-Jun Lai Department of Mathematics University of

More information

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis

ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis Lecture 7: Matrix completion Yuejie Chi The Ohio State University Page 1 Reference Guaranteed Minimum-Rank Solutions of Linear

More information

MATH 532: Linear Algebra

MATH 532: Linear Algebra MATH 532: Linear Algebra Chapter 5: Norms, Inner Products and Orthogonality Greg Fasshauer Department of Applied Mathematics Illinois Institute of Technology Spring 2015 fasshauer@iit.edu MATH 532 1 Outline

More information