Dimensionality reduction: Johnson-Lindenstrauss lemma for structured random matrices
|
|
- Anissa Hancock
- 5 years ago
- Views:
Transcription
1 Dimensionality reduction: Johnson-Lindenstrauss lemma for structured random matrices Jan Vybíral Austrian Academy of Sciences RICAM, Linz, Austria January 2011 MPI Leipzig, Germany
2 joint work with Aicke Hinrichs (University of Jena, Germany) Massimo Fornasier (RICAM, Linz) Jan Haškovec (RICAM, Linz) FWF-START project: Sparse Approximation and Optimization in High Dimensions led by Massimo Fornasier
3 Outline Johnson-Lindenstrauss lemma Classical proof Variants and improvements Circulant matrices Decoupling vs. Fourier transform Applications - Approximate nearest neighbours - Dynamical systems
4 Johnson-Lindenstrauss lemma Let ε (0, 1 2 ), x 1,...,x n R d...arbitrary points, k = O(ε 2 log n), i.e. k Cε 2 log n. There exists a (linear) mapping f : R d R k such that (1 ε) x i x j 2 2 f (x i) f (x j ) 2 2 (1 + ε) x i x j 2 2 for all i,j {1,...,n}. Here 2 stands for the Euclidean norm in R d or R k, respectively.
5 Typical proof A R k d - k d matrices, P - probability measure on A For each y, y 2 = 1: concentration of measure Choosing ( ) P(A A : Ay 2 2 > 1 + ε) exp( ckε2 ), ( ) P(A A : Ay 2 2 < 1 ε) exp( ckε 2 ). exp( ckε 2 ) 1 n 2, the probability of failure (union bound) is smaller then ( ) n n 2 = 1 1 n < 1. Hence, the probability of success is positive!
6 The condition exp( ckε 2 ) 1 n 2 leads to and ckε 2 2log n k 2 c ε 2 log n, i.e. C = 2 c. By increasing C(= 3/c), we may achieve, that such a mapping becomes typical, i.e. occurs with probability at least 1 1/n.
7 Proof of ( ) P...rotational invariant U R d d unitary; X, Y A: P(XU) = P(Y) X := {A : Ay 2 2 > 1 + ε}, Y := {B : Be > 1 + ε}, Ue 1 = y a 1,1... a 1,d A = 1 a 2,1... a 2,d k..... a k,1... a k,d a i,j - indep. Gaussian variables (P is a tensor product) k Be = 1 k i=1 b 2 i,1
8 ( k ) P σi 2 > k(1 + ε) exp( ckε 2 )? i=1 ( k ) ( P σi 2 > k(1 + ε) = P λ i=1 = P ( E exp exp ( ( λ λ ) k σi 2 kλ(1 + ε) > 0 i=1 ) ) k σi 2 kλ(1 + ε) > 1 i=1 ) k σi 2 kλ(1 + ε) i=1 ( 1 = e kλ(1+ε) 2π = e kλ(1+ε) (1 2λ) k/2 R ) k e λt2 e t2 /2 dt
9 Optimization of λ: 0 < λ = 1/2(1 1 1+ε ) < 1/2 leads to e kε/2 (1 + ε) k/2 exp( k/2(ε 2 /2 ε 3 /3))...for 0 < ε < 1/2... exp( ckε 2 )
10 Classical proof W. B. Johnson and J. Lindenstrauss, Extensions of Lipschitz mappings into a Hilbert space. Contem. Math., 26: , 1984 Projection onto a random k-dimensional subspace satisfies the desired property with positive probability advantages: geometrical proof disadvantages: measure on the set of all k-dimensional subspaces evaluating f (x) involves orthonormalisation time consuming
11 Variants and improvements Elementary proof: S. Dasgupta and A. Gupta An elementary proof of a theorem of Johnson and Lindenstrauss. Random. Struct. Algorithms, 22:60-65, Improvements motivated by applications: Good running times of f (x) Small randomness used Small memory space used...others...
12 D. Achlioptas, Database-friendly random projections: Johnson-Lindenstrauss with binary coins. J. Comput. Syst. Sci., 66(4): , f realized by a k d matrix, where each entry is generated independently at random: Gaussian or Bernoulli (or similar) variables. Running time: k d Randomness: k d Memory space k d
13 N. Ailon and B. Chazelle, Approximate nearest neighbors and the fast Johnson-Lindenstrauss transform. In Proc. 38th Annual ACM Symposium on Theory of Computing, f (x) = PHDx, where P is a k d matrix, where each component is generated independently at random P i,j = N(0,1) with probability { q = min Θ P i,j = 0 with probability 1 q, ( log 2 n d ) },1 H is the d d normalized Hadamard matrix, D is a random d d diagonal matrix, with each D i,i drawn independently from { 1,1} with probability 1/2.
14 Running time: With high probability, f (x) may be calculated in time O(d log d + qdk) Randomness: O(d + k log 2 n) Memory space: with high probability O(d +kdq) = O(d +k log 2 n) Not easy to implement Other variants and improvements...
15 Connection to compressed sensing - RIP We say, that A R k d satisfies the Restricted Isometry Property of order s, if there exists δ s (0,1), such that (1 δ s ) x 2 2 Ax 2 2 (1 + δ s) x 2 2 holds for all x R d with x 0 := #{j = 1,...,d : x j 0} s. The aim is to find matrices with small δ s for large s.
16 R. Baraniuk, M. Davenport, R. DeVore and M. Wakin A simple proof of the Restricted Isometry Property for Random Matrices, Constructive Approximation, If P ω ( Aω x 2 2 (1 + ε) x 2 2) exp( nc(ε)) (and the same for ), then A ω satisfies RIP of order s and δ s for s small enough with probability depending on s and δ s. Every distribution that yields J-L transforms, yields also RIP-matrices.
17 Reverse direction provided recently by F. Krahmer and R. Ward: New and improved Johnson-Lindenstrauss embeddings via the Restricted Isometry Property If Φ R k d satisfies RIP of order s (large enough), then ΦD ξ satisfies J-L Lemma. Every distribution that yields RIP matrices, yields also J-L-transforms. Here, D ξ R d d is a diagonal matrix with indep. Bernoulli variables on the diagonal.
18 Circulant matrices a = (a 0,...,a d 1 ) be i.i.d. random variables a 0 a 1 a 2... a d 1 a d 1 a 0 a 1... a d 2 M a,k = a d 2 a d 1 a 0... a d 3 R k d a d k+1 a d k+2 a d k+3... a d k Is it possible to take f (x) = 1 k M a,k x? Or f (x) = 1 k M a,k D κ x?
19 Decoupling vs. Fourier transform Yes! With k = O(ε 2 log 3 n) - decoupling techniques Yes! With k = O(ε 2 log 2 n) - Fourier-analytic methods The improvement to O(ε 2 log n) is still open... promising numerical experiments advantages: running time O(d log d) - using FFT randomness used 2d instead of (k + 1)d easy to implement: FFT is a part of every software package disadvantage: up to now - bigger k
20 log 3 n: Let x 1,...,x n be arbitrary points in R d, ε (0, 1 2 ), k = O(ε 2 log 3 n), a = (a 0,...,a d 1 ) be independent Bernoulli variables or independent normally distributed variables, M a,k and D κ be as above and f (x) = 1 k M a,k D κ x. Then with probability at least 2/3 the following holds (1 ε) x i x j 2 2 f (x i ) f (x j ) 2 2 (1+ε) x i x j 2 2, i,j = 1,...,n.
21 Strategy of the proof of log 3 n -decoupling the dependence Concentration inequalities for every fixed x with x 2 = 1: P a,κ ( M a,k D κ x 2 2 (1 + ε)k ) exp( c(kε 2 ) 1/3 ) and P a,κ ( M a,k D κ x 2 2 (1 ε)k ) exp( c(kε 2 ) 1/3 ). Then union bound over all n(n 1)/2 pairs of points. The bound on k is given by 2 n(n 1) 2 exp( c(kε 2 ) 1/3 ) < 1.
22 Separation of the diagonal and the off-diagonal term k 1 M a,k D κ x 2 2 = d 1 k 1 I = ai 2 i=0 j=0 x 2 j+i } {{ } diagonal j=0 ( d 1 ) 2 a i κ j+i x j+i = I + II i=0 k 1, II = a i a i κ j+i κ j+i x j+i x j+i j=0 i i } {{} off diagonal...summation in the index is modulo d... ) P a,κ ( M a,k D κ x 2 2 (1+ε)k P a (I (1+ε/2)k)+P a,κ (II εk/2)
23 Estimates of I: P a (I (1 + ε/2)k) Lemma of B. Laurent and P. Massart...or any other variant of Bernstein s inequality Exponential concentration of Z = D α i (ai 2 1), i=1 where a i are i.i.d. normal variables and α i are nonnegative real numbers. Then for any t > 0 P(Z 2 α 2 t + 2 α t) exp( t), P(Z 2 α 2 t) exp( t). α i := k 1 j=0 x2 j+i, α 1 = k, α 1 and α 2 k.
24 Estimates of II: Decoupling lemma of Bourgain and Tzafriri: Let ξ 0,...,ξ d 1 be independent random variables with E ξ 0 = = E ξ d 1 = 0 and let {x i,j } d 1 i,j=0 be a double sequence of real numbers. Then for 1 p < E p x i,j ξ i ξ j 4 p E x i,j ξ i ξ j p, i j where (ξ 0,...,ξ d 1 ) denotes an independent copy of (ξ 0,...,ξ d 1 ). i j
25 Further tools: Two times Khintchine s inequalities and ( E a,a k 1 a j a j p) 1/p p(k + p), j=0 for both a Bernoulli or Gaussian variables.
26 The role of D κ k d, a 0,...,a d 1 independent normal variables 2-stability: x = 1 d (1,...,1), b := d 1 j=0 P a ( M a,k x 2 2 > (1 + ε)k depends neither on k nor on d (d 1 M a,k x 2 2 = k a ) j 2 d a j d N(0,1) j=0 ) ( ) = P b b 2 > (1 + ε)
27 log 2 n Let x 1,...,x n be arbitrary points in R d, ε (0, 1 2 ), k = O(ε 2 log 2 n), a = (a 0,...,a d 1 ) be independent normally distributed variables, M a,k and D κ be as above and f (x) = 1 k M a,k D κ x. Then with probability at least 2/3 the following holds (1 ε) x i x j 2 2 f (x i ) f (x j ) 2 2 (1+ε) x i x j 2 2, i,j = 1,...,n.
28 Fourier methods F - unitary discrete Fourier transform, F : C d C d Every circulant matrix may be diagonalised by F and F 1 M a,d x = Fdiag( dfa)f 1 x. The singular values are the square roots of the eigenvalues of M a,d M a,d = Fdiag( dfa)diag( dfa)f 1 = Fdiag(d Fa 2 )F 1 i.e. d Fa.
29 Strategy of the proof of log 2 n Concentration inequalities for all x = x i x j x i x j 2 ( P a Ma,k D κ x 2 2 2(1 + ε)k ) exp ( ckε2 ), log n ( P a Ma,k D κ x 2 2 2(1 ε)k ) exp ( ckε2 ). log n From this, the result follows again by a union bound.
30 Let x 2 = 1. y j := S j (D κ x) C d, j = 0,...,k 1, where S is the shift operator S : C d C d, S(z 0,...,z d 1 ) = (z 1,...,z d 1,z 0 ). Y...k d matrix with rows y 0,...,y k 1. Note, that M a,k D κ x 2 2 = Ya 2 2 Hence, P ( M a,k D κ x 2 2 2(1 + ε)k) = P ( Ya 2 2 2(1 + ε)k).
31 Let Y = UΣV be the singular value decomposition of Y. Then Ya 2 2 = UΣVa 2 2 = ΣVa 2 2 = Σb 2 2, where b := Va is a k-dimensional vector of independent normal variables. Hence, P ( M a,k D κ x 2 2 (1 + ε)k ) = P (k 1 λ 2 j bj 2 (1 + ε)k ), where λ j are the singular values of Y. Lemma of B. Laurent and P. Massart: Estimate λ 4 and λ! j=0 λ 2 2 = Y F = k and λ 2 c log n implies λ 4 4 c k log n.
32 Related results: N. Ailon and E. Liberty, Almost optimal unrestricted fast Johnson-Lindenstrauss transform, Random partial Fourier transform, k = O(ε 4 log n log 4 d) H. Rauhut, J. Romberg and J. Tropp, Restricted isometries for partial random circulant matrices, Random circulant matrices { } k max ε 1 log 3/2 n log 3/2 d,ε 2 log n log 4 d.
33 Approximate nearest neighbors The nearest neighbor problem: Given P = {x 1,...,x n } in a metric space X, preprocess P so as to efficiently find the minimizer of min d(x i,q), q X. i=1,...,n Donald Knuth in vol. 3 of The Art of Computer Programming called it the post-office problem Naive algorithm: compare all the distances - no preprocessing. The approximate nearest neighbor problem: Given P = {x 1,...,x n } in a metric space X and ε > 0, preprocess P so as to efficiently find p P, such that d(p,q) (1 + ε)d(p,q), p X.
34 P = {x 1,...,x n } X = R d, hashing functions: Choose randomly v R d and pre-compute h v (i) := v,x i, i = 1,...,n Find argmin i=1,...,n v,x i q. Iterate over different v 1,v 2,...
35 Dynamical systems J-L-transforms are linear and (almost) preserve distances. We consider non-linear dynamical systems, where the non-linearity depends only on distances of the agents: ẋ i (t) = f i (Dx(t)) + N f ij (Dx(t))x j (t) j=1 N N - number of agents, x(t) = (x 1 (t),...,x N (t)) R d N, where x i : [0,T] R d,i = 1,...,N, f i : R N N R d,i = 1,...,N, f ij : R N N R,i,j = 1,...,N, D : R d N R N N, Dx := ( x i x j l d 2 ),i,j = 1,...,N is the adjacency matrix.
36 Euler s method: x i (0) = x 0 (i),i = 1,...,N and N := xi n + h f i (Dx n ) + f ij (Dx n )xj n,n = 0,...,n 0 1. x n+1 i j=1 We project the system using J-L-transform M R k d : y i (0) = Mx 0 (i),i = 1,...,N and N := yi n +h Mf i (D y n ) + y n+1 i j=1 Mf ij (D y n )y n j If f i and f i,j are Lipschitz: error estimates for y n Mx n.,n = 0,...,n 0 1.
37 References: W. B. Johnson and J. Lindenstrauss, Extensions of Lipschitz mappings into a Hilbert space. Contem. Math., 26: , 1984 S. Dasgupta and A. Gupta, An elementary proof of a theorem of Johnson and Lindenstrauss. Random Struct. & Algorithms, 22:60-65, N. Ailon and B. Chazelle, Approximate nearest neighbors and the fast Johnson-Lindenstrauss transform. In Proc. 38th Annual ACM Symposium on Theory of Computing, A. Hinrichs and J. Vybíral, Johnson-Lindenstrauss lemma for circulant matrices, to appear in Random Struct. & Algorithms J. Vybíral, A variant of the Johnson-Lindenstrauss lemma for circulant matrices, to appear in J. Funct. Anal.
Some Useful Background for Talk on the Fast Johnson-Lindenstrauss Transform
Some Useful Background for Talk on the Fast Johnson-Lindenstrauss Transform Nir Ailon May 22, 2007 This writeup includes very basic background material for the talk on the Fast Johnson Lindenstrauss Transform
More informationRandomized Algorithms
Randomized Algorithms Saniv Kumar, Google Research, NY EECS-6898, Columbia University - Fall, 010 Saniv Kumar 9/13/010 EECS6898 Large Scale Machine Learning 1 Curse of Dimensionality Gaussian Mixture Models
More informationNon-Asymptotic Theory of Random Matrices Lecture 4: Dimension Reduction Date: January 16, 2007
Non-Asymptotic Theory of Random Matrices Lecture 4: Dimension Reduction Date: January 16, 2007 Lecturer: Roman Vershynin Scribe: Matthew Herman 1 Introduction Consider the set X = {n points in R N } where
More informationFaster Johnson-Lindenstrauss style reductions
Faster Johnson-Lindenstrauss style reductions Aditya Menon August 23, 2007 Outline 1 Introduction Dimensionality reduction The Johnson-Lindenstrauss Lemma Speeding up computation 2 The Fast Johnson-Lindenstrauss
More informationFast Random Projections
Fast Random Projections Edo Liberty 1 September 18, 2007 1 Yale University, New Haven CT, supported by AFOSR and NGA (www.edoliberty.com) Advised by Steven Zucker. About This talk will survey a few random
More informationMethods for sparse analysis of high-dimensional data, II
Methods for sparse analysis of high-dimensional data, II Rachel Ward May 23, 2011 High dimensional data with low-dimensional structure 300 by 300 pixel images = 90, 000 dimensions 2 / 47 High dimensional
More informationFast Dimension Reduction
Fast Dimension Reduction Nir Ailon 1 Edo Liberty 2 1 Google Research 2 Yale University Introduction Lemma (Johnson, Lindenstrauss (1984)) A random projection Ψ preserves all ( n 2) distances up to distortion
More informationSparse Johnson-Lindenstrauss Transforms
Sparse Johnson-Lindenstrauss Transforms Jelani Nelson MIT May 24, 211 joint work with Daniel Kane (Harvard) Metric Johnson-Lindenstrauss lemma Metric JL (MJL) Lemma, 1984 Every set of n points in Euclidean
More informationMAT 585: Johnson-Lindenstrauss, Group testing, and Compressed Sensing
MAT 585: Johnson-Lindenstrauss, Group testing, and Compressed Sensing Afonso S. Bandeira April 9, 2015 1 The Johnson-Lindenstrauss Lemma Suppose one has n points, X = {x 1,..., x n }, in R d with d very
More informationMethods for sparse analysis of high-dimensional data, II
Methods for sparse analysis of high-dimensional data, II Rachel Ward May 26, 2011 High dimensional data with low-dimensional structure 300 by 300 pixel images = 90, 000 dimensions 2 / 55 High dimensional
More informationRandom Methods for Linear Algebra
Gittens gittens@acm.caltech.edu Applied and Computational Mathematics California Institue of Technology October 2, 2009 Outline The Johnson-Lindenstrauss Transform 1 The Johnson-Lindenstrauss Transform
More informationFast Dimension Reduction
Fast Dimension Reduction MMDS 2008 Nir Ailon Google Research NY Fast Dimension Reduction Using Rademacher Series on Dual BCH Codes (with Edo Liberty) The Fast Johnson Lindenstrauss Transform (with Bernard
More informationRandom projections. 1 Introduction. 2 Dimensionality reduction. Lecture notes 5 February 29, 2016
Lecture notes 5 February 9, 016 1 Introduction Random projections Random projections are a useful tool in the analysis and processing of high-dimensional data. We will analyze two applications that use
More informationDimensionality Reduction Notes 3
Dimensionality Reduction Notes 3 Jelani Nelson minilek@seas.harvard.edu August 13, 2015 1 Gordon s theorem Let T be a finite subset of some normed vector space with norm X. We say that a sequence T 0 T
More information18.409: Topics in TCS: Embeddings of Finite Metric Spaces. Lecture 6
Massachusetts Institute of Technology Lecturer: Michel X. Goemans 8.409: Topics in TCS: Embeddings of Finite Metric Spaces September 7, 006 Scribe: Benjamin Rossman Lecture 6 Today we loo at dimension
More informationSparser Johnson-Lindenstrauss Transforms
Sparser Johnson-Lindenstrauss Transforms Jelani Nelson Princeton February 16, 212 joint work with Daniel Kane (Stanford) Random Projections x R d, d huge store y = Sx, where S is a k d matrix (compression)
More informationLecture 3. Random Fourier measurements
Lecture 3. Random Fourier measurements 1 Sampling from Fourier matrices 2 Law of Large Numbers and its operator-valued versions 3 Frames. Rudelson s Selection Theorem Sampling from Fourier matrices Our
More information16 Embeddings of the Euclidean metric
16 Embeddings of the Euclidean metric In today s lecture, we will consider how well we can embed n points in the Euclidean metric (l 2 ) into other l p metrics. More formally, we ask the following question.
More informationAcceleration of Randomized Kaczmarz Method
Acceleration of Randomized Kaczmarz Method Deanna Needell [Joint work with Y. Eldar] Stanford University BIRS Banff, March 2011 Problem Background Setup Setup Let Ax = b be an overdetermined consistent
More informationRandomized Algorithms
Randomized Algorithms 南京大学 尹一通 Martingales Definition: A sequence of random variables X 0, X 1,... is a martingale if for all i > 0, E[X i X 0,...,X i1 ] = X i1 x 0, x 1,...,x i1, E[X i X 0 = x 0, X 1
More informationA fast randomized algorithm for overdetermined linear least-squares regression
A fast randomized algorithm for overdetermined linear least-squares regression Vladimir Rokhlin and Mark Tygert Technical Report YALEU/DCS/TR-1403 April 28, 2008 Abstract We introduce a randomized algorithm
More informationOn positive positive-definite functions and Bochner s Theorem
On positive positive-definite functions and Bochner s Theorem Aicke Hinrichs Mathematisches Institut, Universität Jena Ernst-Abbe-Platz, 07740 Jena, Germany email: a.hinrichs@uni-jena.de Jan Vybíral RICAM,
More informationYale university technical report #1402.
The Mailman algorithm: a note on matrix vector multiplication Yale university technical report #1402. Edo Liberty Computer Science Yale University New Haven, CT Steven W. Zucker Computer Science and Appled
More informationAccelerated Dense Random Projections
1 Advisor: Steven Zucker 1 Yale University, Department of Computer Science. Dimensionality reduction (1 ε) xi x j 2 Ψ(xi ) Ψ(x j ) 2 (1 + ε) xi x j 2 ( n 2) distances are ε preserved Target dimension k
More informationCS on CS: Computer Science insights into Compresive Sensing (and vice versa) Piotr Indyk MIT
CS on CS: Computer Science insights into Compresive Sensing (and vice versa) Piotr Indyk MIT Sparse Approximations Goal: approximate a highdimensional vector x by x that is sparse, i.e., has few nonzero
More informationMultivariate Statistics Random Projections and Johnson-Lindenstrauss Lemma
Multivariate Statistics Random Projections and Johnson-Lindenstrauss Lemma Suppose again we have n sample points x,..., x n R p. The data-point x i R p can be thought of as the i-th row X i of an n p-dimensional
More informationSupremum of simple stochastic processes
Subspace embeddings Daniel Hsu COMS 4772 1 Supremum of simple stochastic processes 2 Recap: JL lemma JL lemma. For any ε (0, 1/2), point set S R d of cardinality 16 ln n S = n, and k N such that k, there
More informationJOHNSON-LINDENSTRAUSS TRANSFORMATION AND RANDOM PROJECTION
JOHNSON-LINDENSTRAUSS TRANSFORMATION AND RANDOM PROJECTION LONG CHEN ABSTRACT. We give a brief survey of Johnson-Lindenstrauss lemma. CONTENTS 1. Introduction 1 2. JL Transform 4 2.1. An Elementary Proof
More informationOptimal compression of approximate Euclidean distances
Optimal compression of approximate Euclidean distances Noga Alon 1 Bo az Klartag 2 Abstract Let X be a set of n points of norm at most 1 in the Euclidean space R k, and suppose ε > 0. An ε-distance sketch
More informationLimitations in Approximating RIP
Alok Puranik Mentor: Adrian Vladu Fifth Annual PRIMES Conference, 2015 Outline 1 Background The Problem Motivation Construction Certification 2 Planted model Planting eigenvalues Analysis Distinguishing
More informationKernel Method: Data Analysis with Positive Definite Kernels
Kernel Method: Data Analysis with Positive Definite Kernels 2. Positive Definite Kernel and Reproducing Kernel Hilbert Space Kenji Fukumizu The Institute of Statistical Mathematics. Graduate University
More informationCS 229r: Algorithms for Big Data Fall Lecture 17 10/28
CS 229r: Algorithms for Big Data Fall 2015 Prof. Jelani Nelson Lecture 17 10/28 Scribe: Morris Yau 1 Overview In the last lecture we defined subspace embeddings a subspace embedding is a linear transformation
More informationUsing the Johnson-Lindenstrauss lemma in linear and integer programming
Using the Johnson-Lindenstrauss lemma in linear and integer programming Vu Khac Ky 1, Pierre-Louis Poirion, Leo Liberti LIX, École Polytechnique, F-91128 Palaiseau, France Email:{vu,poirion,liberti}@lix.polytechnique.fr
More informationAdvances in Manifold Learning Presented by: Naku Nak l Verm r a June 10, 2008
Advances in Manifold Learning Presented by: Nakul Verma June 10, 008 Outline Motivation Manifolds Manifold Learning Random projection of manifolds for dimension reduction Introduction to random projections
More informationVery Sparse Random Projections
Very Sparse Random Projections Ping Li, Trevor Hastie and Kenneth Church [KDD 06] Presented by: Aditya Menon UCSD March 4, 2009 Presented by: Aditya Menon (UCSD) Very Sparse Random Projections March 4,
More informationLecture 13 October 6, Covering Numbers and Maurey s Empirical Method
CS 395T: Sublinear Algorithms Fall 2016 Prof. Eric Price Lecture 13 October 6, 2016 Scribe: Kiyeon Jeon and Loc Hoang 1 Overview In the last lecture we covered the lower bound for p th moment (p > 2) and
More informationCompressed Sensing: Lecture I. Ronald DeVore
Compressed Sensing: Lecture I Ronald DeVore Motivation Compressed Sensing is a new paradigm for signal/image/function acquisition Motivation Compressed Sensing is a new paradigm for signal/image/function
More informationDense Fast Random Projections and Lean Walsh Transforms
Dense Fast Random Projections and Lean Walsh Transforms Edo Liberty, Nir Ailon, and Amit Singer Abstract. Random projection methods give distributions over k d matrices such that if a matrix Ψ (chosen
More informationFast Random Projections using Lean Walsh Transforms Yale University Technical report #1390
Fast Random Projections using Lean Walsh Transforms Yale University Technical report #1390 Edo Liberty Nir Ailon Amit Singer Abstract We present a k d random projection matrix that is applicable to vectors
More informationA Simple Proof of the Restricted Isometry Property for Random Matrices
DOI 10.1007/s00365-007-9003-x A Simple Proof of the Restricted Isometry Property for Random Matrices Richard Baraniuk Mark Davenport Ronald DeVore Michael Wakin Received: 17 May 006 / Revised: 18 January
More informationGuaranteed Minimum-Rank Solutions of Linear Matrix Equations via Nuclear Norm Minimization
Forty-Fifth Annual Allerton Conference Allerton House, UIUC, Illinois, USA September 26-28, 27 WeA3.2 Guaranteed Minimum-Rank Solutions of Linear Matrix Equations via Nuclear Norm Minimization Benjamin
More informationReconstruction from Anisotropic Random Measurements
Reconstruction from Anisotropic Random Measurements Mark Rudelson and Shuheng Zhou The University of Michigan, Ann Arbor Coding, Complexity, and Sparsity Workshop, 013 Ann Arbor, Michigan August 7, 013
More informationImproved Bounds on the Dot Product under Random Projection and Random Sign Projection
Improved Bounds on the Dot Product under Random Projection and Random Sign Projection Ata Kabán School of Computer Science The University of Birmingham Birmingham B15 2TT, UK http://www.cs.bham.ac.uk/
More informationStrengthened Sobolev inequalities for a random subspace of functions
Strengthened Sobolev inequalities for a random subspace of functions Rachel Ward University of Texas at Austin April 2013 2 Discrete Sobolev inequalities Proposition (Sobolev inequality for discrete images)
More informationNear Ideal Behavior of a Modified Elastic Net Algorithm in Compressed Sensing
Near Ideal Behavior of a Modified Elastic Net Algorithm in Compressed Sensing M. Vidyasagar Cecil & Ida Green Chair The University of Texas at Dallas M.Vidyasagar@utdallas.edu www.utdallas.edu/ m.vidyasagar
More informationLecture 18: March 15
CS71 Randomness & Computation Spring 018 Instructor: Alistair Sinclair Lecture 18: March 15 Disclaimer: These notes have not been subjected to the usual scrutiny accorded to formal publications. They may
More information1 Dimension Reduction in Euclidean Space
CSIS0351/8601: Randomized Algorithms Lecture 6: Johnson-Lindenstrauss Lemma: Dimension Reduction Lecturer: Hubert Chan Date: 10 Oct 011 hese lecture notes are supplementary materials for the lectures.
More informationThe Johnson-Lindenstrauss Lemma
The Johnson-Lindenstrauss Lemma Kevin (Min Seong) Park MAT477 Introduction The Johnson-Lindenstrauss Lemma was first introduced in the paper Extensions of Lipschitz mappings into a Hilbert Space by William
More informationExponential decay of reconstruction error from binary measurements of sparse signals
Exponential decay of reconstruction error from binary measurements of sparse signals Deanna Needell Joint work with R. Baraniuk, S. Foucart, Y. Plan, and M. Wootters Outline Introduction Mathematical Formulation
More informationA fast randomized algorithm for approximating an SVD of a matrix
A fast randomized algorithm for approximating an SVD of a matrix Joint work with Franco Woolfe, Edo Liberty, and Vladimir Rokhlin Mark Tygert Program in Applied Mathematics Yale University Place July 17,
More informationsample lectures: Compressed Sensing, Random Sampling I & II
P. Jung, CommIT, TU-Berlin]: sample lectures Compressed Sensing / Random Sampling I & II 1/68 sample lectures: Compressed Sensing, Random Sampling I & II Peter Jung Communications and Information Theory
More informationCS168: The Modern Algorithmic Toolbox Lecture #4: Dimensionality Reduction
CS168: The Modern Algorithmic Toolbox Lecture #4: Dimensionality Reduction Tim Roughgarden & Gregory Valiant April 12, 2017 1 The Curse of Dimensionality in the Nearest Neighbor Problem Lectures #1 and
More informationJohnson-Lindenstrauss, Concentration and applications to Support Vector Machines and Kernels
Johnson-Lindenstrauss, Concentration and applications to Support Vector Machines and Kernels Devdatt Dubhashi Department of Computer Science and Engineering, Chalmers University, dubhashi@chalmers.se Functional
More informationarxiv: v1 [math.pr] 22 May 2008
THE LEAST SINGULAR VALUE OF A RANDOM SQUARE MATRIX IS O(n 1/2 ) arxiv:0805.3407v1 [math.pr] 22 May 2008 MARK RUDELSON AND ROMAN VERSHYNIN Abstract. Let A be a matrix whose entries are real i.i.d. centered
More informationSketching as a Tool for Numerical Linear Algebra
Sketching as a Tool for Numerical Linear Algebra David P. Woodruff presented by Sepehr Assadi o(n) Big Data Reading Group University of Pennsylvania February, 2015 Sepehr Assadi (Penn) Sketching for Numerical
More informationNearest Neighbor Preserving Embeddings
Nearest Neighbor Preserving Embeddings Piotr Indyk MIT Assaf Naor Microsoft Research Abstract In this paper we introduce the notion of nearest neighbor preserving embeddings. These are randomized embeddings
More informationSTAT 200C: High-dimensional Statistics
STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 59 Classical case: n d. Asymptotic assumption: d is fixed and n. Basic tools: LLN and CLT. High-dimensional setting: n d, e.g. n/d
More informationAN INTRODUCTION TO COMPRESSIVE SENSING
AN INTRODUCTION TO COMPRESSIVE SENSING Rodrigo B. Platte School of Mathematical and Statistical Sciences APM/EEE598 Reverse Engineering of Complex Dynamical Networks OUTLINE 1 INTRODUCTION 2 INCOHERENCE
More informationConstructing Explicit RIP Matrices and the Square-Root Bottleneck
Constructing Explicit RIP Matrices and the Square-Root Bottleneck Ryan Cinoman July 18, 2018 Ryan Cinoman Constructing Explicit RIP Matrices July 18, 2018 1 / 36 Outline 1 Introduction 2 Restricted Isometry
More informationInvertibility of symmetric random matrices
Invertibility of symmetric random matrices Roman Vershynin University of Michigan romanv@umich.edu February 1, 2011; last revised March 16, 2012 Abstract We study n n symmetric random matrices H, possibly
More informationElements of Positive Definite Kernel and Reproducing Kernel Hilbert Space
Elements of Positive Definite Kernel and Reproducing Kernel Hilbert Space Statistical Inference with Reproducing Kernel Hilbert Space Kenji Fukumizu Institute of Statistical Mathematics, ROIS Department
More informationRandom regular digraphs: singularity and spectrum
Random regular digraphs: singularity and spectrum Nick Cook, UCLA Probability Seminar, Stanford University November 2, 2015 Universality Circular law Singularity probability Talk outline 1 Universality
More informationCompressibility of Infinite Sequences and its Interplay with Compressed Sensing Recovery
Compressibility of Infinite Sequences and its Interplay with Compressed Sensing Recovery Jorge F. Silva and Eduardo Pavez Department of Electrical Engineering Information and Decision Systems Group Universidad
More informationSketching as a Tool for Numerical Linear Algebra All Lectures. David Woodruff IBM Almaden
Sketching as a Tool for Numerical Linear Algebra All Lectures David Woodruff IBM Almaden Massive data sets Examples Internet traffic logs Financial data etc. Algorithms Want nearly linear time or less
More informationReview of similarity transformation and Singular Value Decomposition
Review of similarity transformation and Singular Value Decomposition Nasser M Abbasi Applied Mathematics Department, California State University, Fullerton July 8 7 page compiled on June 9, 5 at 9:5pm
More informationInvertibility of random matrices
University of Michigan February 2011, Princeton University Origins of Random Matrix Theory Statistics (Wishart matrices) PCA of a multivariate Gaussian distribution. [Gaël Varoquaux s blog gael-varoquaux.info]
More informationRandomness Efficient Fast-Johnson-Lindenstrauss Transform with Applications in Differential Privacy
Randomness Efficient Fast-Johnson-Lindenstrauss Transform with Applications in Differential Privacy Jalaj Upadhyay Cheriton School of Computer Science University of Waterloo jkupadhy@csuwaterlooca arxiv:14102470v6
More informationThe Singular Value Decomposition
The Singular Value Decomposition Philippe B. Laval KSU Fall 2015 Philippe B. Laval (KSU) SVD Fall 2015 1 / 13 Review of Key Concepts We review some key definitions and results about matrices that will
More informationModel Counting for Logical Theories
Model Counting for Logical Theories Wednesday Dmitry Chistikov Rayna Dimitrova Department of Computer Science University of Oxford, UK Max Planck Institute for Software Systems (MPI-SWS) Kaiserslautern
More informationLinear Algebra Review
January 29, 2013 Table of contents Metrics Metric Given a space X, then d : X X R + 0 and z in X if: d(x, y) = 0 is equivalent to x = y d(x, y) = d(y, x) d(x, y) d(x, z) + d(z, y) is a metric is for all
More informationHomework 1. Yuan Yao. September 18, 2011
Homework 1 Yuan Yao September 18, 2011 1. Singular Value Decomposition: The goal of this exercise is to refresh your memory about the singular value decomposition and matrix norms. A good reference to
More informationLecture Notes 1: Vector spaces
Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector
More informationRandomized Numerical Linear Algebra: Review and Progresses
ized ized SVD ized : Review and Progresses Zhihua Department of Computer Science and Engineering Shanghai Jiao Tong University The 12th China Workshop on Machine Learning and Applications Xi an, November
More informationIntroduction to Compressed Sensing
Introduction to Compressed Sensing Alejandro Parada, Gonzalo Arce University of Delaware August 25, 2016 Motivation: Classical Sampling 1 Motivation: Classical Sampling Issues Some applications Radar Spectral
More informationHomework 2 Foundations of Computational Math 2 Spring 2019
Homework 2 Foundations of Computational Math 2 Spring 2019 Problem 2.1 (2.1.a) Suppose (v 1,λ 1 )and(v 2,λ 2 ) are eigenpairs for a matrix A C n n. Show that if λ 1 λ 2 then v 1 and v 2 are linearly independent.
More informationLecture Notes 9: Constrained Optimization
Optimization-based data analysis Fall 017 Lecture Notes 9: Constrained Optimization 1 Compressed sensing 1.1 Underdetermined linear inverse problems Linear inverse problems model measurements of the form
More informationModel-Based Compressive Sensing for Signal Ensembles. Marco F. Duarte Volkan Cevher Richard G. Baraniuk
Model-Based Compressive Sensing for Signal Ensembles Marco F. Duarte Volkan Cevher Richard G. Baraniuk Concise Signal Structure Sparse signal: only K out of N coordinates nonzero model: union of K-dimensional
More informationA Power Efficient Sensing/Communication Scheme: Joint Source-Channel-Network Coding by Using Compressive Sensing
Forty-Ninth Annual Allerton Conference Allerton House, UIUC, Illinois, USA September 28-30, 20 A Power Efficient Sensing/Communication Scheme: Joint Source-Channel-Network Coding by Using Compressive Sensing
More informationConcentration of Measure Inequalities for. Toeplitz Matrices with Applications
Concentration of Measure Inequalities for 1 Toeplitz Matrices with Applications Borhan M. Sanandaji, Tyrone L. Vincent, and Michael B. Wakin Abstract We derive Concentration of Measure (CoM) inequalities
More informationNotes taken by Costis Georgiou revised by Hamed Hatami
CSC414 - Metric Embeddings Lecture 6: Reductions that preserve volumes and distance to affine spaces & Lower bound techniques for distortion when embedding into l Notes taken by Costis Georgiou revised
More informationSTA141C: Big Data & High Performance Statistical Computing
STA141C: Big Data & High Performance Statistical Computing Numerical Linear Algebra Background Cho-Jui Hsieh UC Davis May 15, 2018 Linear Algebra Background Vectors A vector has a direction and a magnitude
More informationSparse Legendre expansions via l 1 minimization
Sparse Legendre expansions via l 1 minimization Rachel Ward, Courant Institute, NYU Joint work with Holger Rauhut, Hausdorff Center for Mathematics, Bonn, Germany. June 8, 2010 Outline Sparse recovery
More informationApproximate Spectral Clustering via Randomized Sketching
Approximate Spectral Clustering via Randomized Sketching Christos Boutsidis Yahoo! Labs, New York Joint work with Alex Gittens (Ebay), Anju Kambadur (IBM) The big picture: sketch and solve Tradeoff: Speed
More informationLecture 15: Random Projections
Lecture 15: Random Projections Introduction to Learning and Analysis of Big Data Kontorovich and Sabato (BGU) Lecture 15 1 / 11 Review of PCA Unsupervised learning technique Performs dimensionality reduction
More informationA Unified Theorem on SDP Rank Reduction. yyye
SDP Rank Reduction Yinyu Ye, EURO XXII 1 A Unified Theorem on SDP Rank Reduction Yinyu Ye Department of Management Science and Engineering and Institute of Computational and Mathematical Engineering Stanford
More informationThe Johnson-Lindenstrauss Lemma in Linear Programming
The Johnson-Lindenstrauss Lemma in Linear Programming Leo Liberti, Vu Khac Ky, Pierre-Louis Poirion CNRS LIX Ecole Polytechnique, France Aussois COW 2016 The gist Goal: solving very large LPs min{c x Ax
More informationarxiv: v1 [cs.ds] 30 May 2010
ALMOST OPTIMAL UNRESTRICTED FAST JOHNSON-LINDENSTRAUSS TRANSFORM arxiv:005.553v [cs.ds] 30 May 200 NIR AILON AND EDO LIBERTY Abstract. The problems of random projections and sparse reconstruction have
More informationMatrix Theory. A.Holst, V.Ufnarovski
Matrix Theory AHolst, VUfnarovski 55 HINTS AND ANSWERS 9 55 Hints and answers There are two different approaches In the first one write A as a block of rows and note that in B = E ij A all rows different
More informationCS 229r: Algorithms for Big Data Fall Lecture 19 Nov 5
CS 229r: Algorithms for Big Data Fall 215 Prof. Jelani Nelson Lecture 19 Nov 5 Scribe: Abdul Wasay 1 Overview In the last lecture, we started discussing the problem of compressed sensing where we are given
More informationQuantum Computing Lecture 2. Review of Linear Algebra
Quantum Computing Lecture 2 Review of Linear Algebra Maris Ozols Linear algebra States of a quantum system form a vector space and their transformations are described by linear operators Vector spaces
More informationCompressive Sensing and Beyond
Compressive Sensing and Beyond Sohail Bahmani Gerorgia Tech. Signal Processing Compressed Sensing Signal Models Classics: bandlimited The Sampling Theorem Any signal with bandwidth B can be recovered
More informationFast Parallel Estimation of High Dimensional Information Theoretical Quantities with Low Dimensional Random Projection Ensembles
Fast Parallel Estimation of High Dimensional Information Theoretical Quantities with Low Dimensional Random Projection Ensembles Zoltán Szabó and András Lőrincz Department of Information Systems, Eötvös
More informationConstrained optimization
Constrained optimization DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis http://www.cims.nyu.edu/~cfgranda/pages/obda_fall17/index.html Carlos Fernandez-Granda Compressed sensing Convex constrained
More informationTHE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR
THE PERTURBATION BOUND FOR THE SPECTRAL RADIUS OF A NON-NEGATIVE TENSOR WEN LI AND MICHAEL K. NG Abstract. In this paper, we study the perturbation bound for the spectral radius of an m th - order n-dimensional
More informationMANY scientific computations, signal processing, data analysis and machine learning applications lead to large dimensional
Low rank approximation and decomposition of large matrices using error correcting codes Shashanka Ubaru, Arya Mazumdar, and Yousef Saad 1 arxiv:1512.09156v3 [cs.it] 15 Jun 2017 Abstract Low rank approximation
More informationHigh Dimensional Geometry, Curse of Dimensionality, Dimension Reduction
Chapter 11 High Dimensional Geometry, Curse of Dimensionality, Dimension Reduction High-dimensional vectors are ubiquitous in applications (gene expression data, set of movies watched by Netflix customer,
More informationSketching as a Tool for Numerical Linear Algebra
Sketching as a Tool for Numerical Linear Algebra (Part 2) David P. Woodruff presented by Sepehr Assadi o(n) Big Data Reading Group University of Pennsylvania February, 2015 Sepehr Assadi (Penn) Sketching
More informationSparse Recovery with Pre-Gaussian Random Matrices
Sparse Recovery with Pre-Gaussian Random Matrices Simon Foucart Laboratoire Jacques-Louis Lions Université Pierre et Marie Curie Paris, 75013, France Ming-Jun Lai Department of Mathematics University of
More informationECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis
ECE 8201: Low-dimensional Signal Models for High-dimensional Data Analysis Lecture 7: Matrix completion Yuejie Chi The Ohio State University Page 1 Reference Guaranteed Minimum-Rank Solutions of Linear
More informationMATH 532: Linear Algebra
MATH 532: Linear Algebra Chapter 5: Norms, Inner Products and Orthogonality Greg Fasshauer Department of Applied Mathematics Illinois Institute of Technology Spring 2015 fasshauer@iit.edu MATH 532 1 Outline
More information