Non-Asymptotic Theory of Random Matrices Lecture 4: Dimension Reduction Date: January 16, 2007

Similar documents
16 Embeddings of the Euclidean metric

Metric structures in L 1 : Dimension, snowflakes, and average distortion

Lectures in Geometric Functional Analysis. Roman Vershynin

Optimal compression of approximate Euclidean distances

Some Useful Background for Talk on the Fast Johnson-Lindenstrauss Transform

On metric characterizations of some classes of Banach spaces

Dimensionality reduction: Johnson-Lindenstrauss lemma for structured random matrices

Nearest Neighbor Preserving Embeddings

1 Dimension Reduction in Euclidean Space

Empirical Processes and random projections

An Algorithmist s Toolkit Nov. 10, Lecture 17

E.7 Alaoglu s Theorem

18.409: Topics in TCS: Embeddings of Finite Metric Spaces. Lecture 6

Random sets of isomorphism of linear operators on Hilbert space

Methods for sparse analysis of high-dimensional data, II

Isoperimetry of waists and local versus global asymptotic convex geometries

Analysis Comprehensive Exam Questions Fall F(x) = 1 x. f(t)dt. t 1 2. tf 2 (t)dt. and g(t, x) = 2 t. 2 t

Lecture 18: March 15

Lecture Notes in Advanced Calculus 1 (80315) Raz Kupferman Institute of Mathematics The Hebrew University

Z Algorithmic Superpower Randomization October 15th, Lecture 12

arxiv:math/ v1 [math.fa] 30 Sep 2004

Basic Properties of Metric and Normed Spaces

Methods for sparse analysis of high-dimensional data, II

Randomized Algorithms

A Banach space with a symmetric basis which is of weak cotype 2 but not of cotype 2

Randomized Algorithms

Probabilistic Methods in Asymptotic Geometric Analysis.

Chapter 2 Metric Spaces

A Bernstein-Chernoff deviation inequality, and geometric properties of random families of operators

Asymptotic Geometric Analysis, Fall 2006

The small ball property in Banach spaces (quantitative results)

Upper Bound for Intermediate Singular Values of Random Sub-Gaussian Matrices 1

MATH 51H Section 4. October 16, Recall what it means for a function between metric spaces to be continuous:

Ergodic Theory and Topological Groups

An example of a convex body without symmetric projections.

Finite Metric Spaces & Their Embeddings: Introduction and Basic Tools

Math 421, Homework #9 Solutions

Using the Johnson-Lindenstrauss lemma in linear and integer programming

Chapter 14. Duality for Normed Linear Spaces

The Fourier transform and finite differences: an exposition and proof. of a result of Gary H. Meisters and Wolfgang M. Schmidt

Sections of Convex Bodies via the Combinatorial Dimension

THE RADON NIKODYM PROPERTY AND LIPSCHITZ EMBEDDINGS

The covering numbers and low M -estimate for quasi-convex bodies.

Takens embedding theorem for infinite-dimensional dynamical systems

9 Radon-Nikodym theorem and conditioning

Faster Johnson-Lindenstrauss style reductions

The Hopf argument. Yves Coudene. IRMAR, Université Rennes 1, campus beaulieu, bat Rennes cedex, France

Lebesgue Measure on R n

Maths 212: Homework Solutions

Lebesgue Measure on R n

Elements of Convex Optimization Theory

FRACTAL GEOMETRY, DYNAMICAL SYSTEMS AND CHAOS

FUNCTIONAL ANALYSIS LECTURE NOTES: COMPACT SETS AND FINITE-DIMENSIONAL SPACES. 1. Compact Sets

Random walk on groups

Problem Set 6: Solutions Math 201A: Fall a n x n,

NOWHERE LOCALLY UNIFORMLY CONTINUOUS FUNCTIONS

A BOOLEAN ACTION OF C(M, U(1)) WITHOUT A SPATIAL MODEL AND A RE-EXAMINATION OF THE CAMERON MARTIN THEOREM

ON SUBSPACES OF NON-REFLEXIVE ORLICZ SPACES

ON ASSOUAD S EMBEDDING TECHNIQUE

ABSOLUTE CONTINUITY OF FOLIATIONS

Linear Analysis Lecture 5

MATH 202B - Problem Set 5

Small ball probability and Dvoretzky Theorem

Rate of convergence of geometric symmetrizations

Szemerédi s Lemma for the Analyst

MAT 570 REAL ANALYSIS LECTURE NOTES. Contents. 1. Sets Functions Countability Axiom of choice Equivalence relations 9

arxiv: v1 [math.pr] 22 May 2008

René Bartsch and Harry Poppe (Received 4 July, 2015)

Measurable functions are approximately nice, even if look terrible.

THEOREMS, ETC., FOR MATH 515

APPROXIMATE ISOMETRIES ON FINITE-DIMENSIONAL NORMED SPACES

Lecture Notes 1: Vector spaces

Ordinal Embedding: Approximation Algorithms and Dimensionality Reduction

Problem Set 2: Solutions Math 201A: Fall 2016

Star bodies with completely symmetric sections

The Knaster problem and the geometry of high-dimensional cubes

On the Equivalence Between Geometric and Arithmetic Means for Log-Concave Measures

BALANCING GAUSSIAN VECTORS. 1. Introduction

Math 5210, Definitions and Theorems on Metric Spaces

Spectral Gap and Concentration for Some Spherically Symmetric Probability Measures

Diamond graphs and super-reflexivity

Introduction to Real Analysis Alternative Chapter 1

Lecture 5 : Projections

Non-linear factorization of linear operators

Multivariate Statistics Random Projections and Johnson-Lindenstrauss Lemma

A note on the σ-algebra of cylinder sets and all that

Advances in Manifold Learning Presented by: Naku Nak l Verm r a June 10, 2008

Notes taken by Costis Georgiou revised by Hamed Hatami

Small distortion and volume preserving embedding for Planar and Euclidian metrics Satish Rao

KUIPER S THEOREM ON CONFORMALLY FLAT MANIFOLDS

LECTURE NOTES ON DVORETZKY S THEOREM

Fast Dimension Reduction

Analysis II Lecture notes

Doubling metric spaces and embeddings. Assaf Naor

Chapter 3: Baire category and open mapping theorems

Random Methods for Linear Algebra

Random hyperplane tessellations and dimension reduction

QUOTIENTS OF ESSENTIALLY EUCLIDEAN SPACES

SPECTRAL THEOREM FOR COMPACT SELF-ADJOINT OPERATORS

FAT AND THIN SETS FOR DOUBLING MEASURES IN EUCLIDEAN SPACE

Transcription:

Non-Asymptotic Theory of Random Matrices Lecture 4: Dimension Reduction Date: January 16, 2007 Lecturer: Roman Vershynin Scribe: Matthew Herman 1 Introduction Consider the set X = {n points in R N } where both n and N are large. Our goal will be to reduce N so as to represent X in a space of dimension k N. One motivation for this comes from computer science where X is a data structure. Observation 1. 1. We can always take N = n, since dim(span(x)) = n and X span(x). 2. For N < n we may lose linear independence. Goal: We want to construct a map T : R N R k, k << N such that pairwise distances in X are approximately preserved. That is, for all x, y X there exists ε > 0 such that (1 ε) x y 2 T x T y 2 (1 + ε) x y 2. Such a map is called an ε-embedding of our data structure X into R k (more precisely into l k 2 ). For us, T will be the linear map associated with a random matrix. Theorem 2 (Johnson-Lindenstrauss (J-L) Flattening Lemma [3]). Let X be an n-point set in a Hilbert space, and suppose ε (0, 1). Then there exists an ε-embedding of X into l k 2, for some k Cε 2 log n and some constant C > 0. Example 3. To represent n points in R n in a computer we need to store n 2 numbers. However, by the J-L Flattening Lemma we can store only O(n log n) numbers and still reconstruct all distances within ε-error. 1

All known embeddings satisfying Theorem 2 are given by random matrices T such as Gaussian Bernoulli (Achlioptas [1]) Orthogonal projections (J-L) 1.1 Random Projections in R n Random rotations (= random orthogonal matrices) Orthogonal group O n = {rotations in R n } = {orthogonal n n matrices} Here, O n is a probability space with probability measure called Haar Measure (from topological group theory). 1.2 Haar Measure Suppose that M is a compact metric space (such as a sphere in R n ), and that G is a group of isometries of M. Theorem 4 (Haar Measure). 1. There exists a Borel probability measure µ on M which is invariant under G. That is, µ(s) = µ(gs) for all g G and S M. 2. If G is transitive, then the Haar measure is unique. Here, transitive means for all x, y X, there exists g G such that gx = y. Proof: See 1 of Milman-Schechtmann [7] for a simple 2-page proof. Example 5. Let M = S n 1, G = O n. Then µ = usual Lebesgue measure. Example 6. Let M = G = O n. Here the metric on a rotation comes from the Hilbert-Schmidt norm. In this sense n n matrices can be viewed as vectors in R n2, and the Hilbert-Schmidt norm provides a Euclidean distance between these matrices. This gives us a Haar measure on O n. 2

1.3 Random rotations, random orthogonal matrices Fact 7 (How to compute Haar Measure). For a given z S n 1 and for a given random rotation (with respect to Haar measure) U O n, Uz is a random vector which is uniformly distributed on S n 1. Proof. We want to show for all A S n 1 that we have P(Uz A) = σ(a), where P denotes probability (with respect to Haar measure) and σ is the usual uniform measure. Note that both sides of this equation define a probability measure on S n 1 which are rotationally invariant. The uniqueness part of the Haar Measure theorem implies that they coincide. Definition 8. Random k-subspace of R n = random rotation of R k R n, denoted by U (R k ) Random Projection = orthogonal projection onto a random k-subspace, denoted by U P k U, where P k : R n R k is a projection (i.e., P k U = the first k rows of matrix U) Proof of J-L Lemma (for ε-embedding given by random projections) Layout of proof : First we will show that J-L holds for one particular fixed pair x, y X. Then we will show that it holds collectively for all possible pairs in X. Step 1. Fix an arbitrary pair x, y X, and let z = x y. Lemma 9 ( Norm of a random projection). Let z R n be fixed, and let P be a random projection in R n onto a k-subspace. Then 1. (E P z 2 2 )1/2 = k/n, where E denotes expectation 2. For ε > 0 we have that (1 ε) k/n P z 2 (1 + ε) k/n holds with probability 1 2e kε2 /2 Proof. 1. Note that the statement Projecting a fixed z onto a random subspace is equivalent to the statement Projecting a random z onto a fixed subspace. Then with random rotation U we have P z 2 = U P k Uz 2 = P k Uz 2 = P k x 2 3

where x is a random vector distributed uniformly on S n 1. Then E P k x 2 2 = E k x 2 j = j=1 k E x 2 j = k E x 2 1 = k/n. j=1 2. For the next part of Lemma 9 we want to show P Failure Probability P( z 2 k/n > ε ) k/n ( = σ {x S n 1 : P k x 2 k/n > ε ) k/n}. Let f(x) = P k x 2, and recognize that f : S n 1 R n is a 1-Lipschitz function. From Part 1 of this lemma we have that ( Ef 2) 1/2 = k/n. Let p = σ({x S n 1 : f(x) (Ef 2 ) 1/2 > ε k/n}). Then p 2e n(ε k/n ) 2 /2 = 2e kε2 /2 where we used the results of Concentration of Measure discussed in 2 of Lecture 3. We next normalize our embedding as T := n/k P. By Lemma 9, we have for any fixed z X that (1 ε) z 2 T z 2 (1 + ε) z 2 (1) holds with high probability. Thus for any fixed x, y X and putting z = x y we have that the inequality (1) holds with probability 1 2e kε2 /2. Now we consider the collective ensemble. Notice that there are no more than n 2 pairs of x, y X. Then taking the union over all pairs we have the probability that (1) fails for some pair is less than or equal to n 2 2e kε2 /2 which is less than 1 if k Cε 2 log n. Hence with positive probability (1) holds with for all pairs. (End of proof for the J-L Flattening Lemma.) 4

Example 10 (Argument that k C(ε) log n using volume point of view). Suppose for all x, y X that 1/2 x y 2 1. Now suppose that ε = 1/10. Then by the J-L Flattening Lemma we can claim that 1/4 (1 ε)/2 T x T y 2 (1 + ε) 2. This tell us the following: The image of X under T is contained in a ball of radius 2 (i.e., T (X) 2B k 2 ) There are n disjoint 1 8 -balls centered at points in T (X) (2+ 1 8 )Bk 2 3B2 k if n Vol( 1 8 Bk 2 ) Vol(3Bk 2 ) n ( 1 8 )k Vol(B k 2 ) 3 k Vol(B k 2 ) n 24 k log n k. The last example showed that the log n factor in the J-L Flattening Lemma is sharp. The sharpness of the ε 2 factor is due to Alon and can be found in [6]. 5

2 Related Work Is there a version for l 1? No: Lee and Naor [4], Brinkman and Charikar [2]. l? No. l p (for p 1, 2, )? Still open. 3 Related Readings Matousek [6] and [5] Vempala [8] References [1] D. Achlioptas. Database-friendly random projections. Proc. 20th Annual Symposium on Principles of Database Systems, 26:274 281, 2001. [2] B. Brinkman and M. Charikar. On the impossibility of dimension reduction in l 1. Journal of ACM, 52(5):766 788, 2005. [3] W. B. Johnson and J. Lindenstrauss. Extensions of lipschitz mappings into a hilbert space. Contemp. Math., 26:189 206, 1984. [4] J. R. Lee and A. Naor. Embedding the diamond graph in l p and dimension reduction in l 1. Geometric and Functional Analysis, 14(4):745 747, 2004. [5] J. Matousek. Website on open problems on embeddings of metric spaces. http://kam.mff.cuni.cz/ matousek/. [6] J. Matousek. Lectures on Discrete Geometry. Springer, Berlin- Heidelberg New York, 2002. [7] Vitali D. Milman and Gideon Schechtman. Asymptotic theory of finitedimensional normed spaces, volume 1200 of Lecture Notes in Mathematics. Springer-Verlag, Berlin, 1986. [8] Santosh S. Vempala. The random projection method, volume 65 of DI- MACS Series in Discrete Mathematics and Theoretical Computer Science. American Mathematical Society, Providence, 2004. 6