Finite Metric Spaces & Their Embeddings: Introduction and Basic Tools

Similar documents
Partitioning Metric Spaces

Topic: Balanced Cut, Sparsest Cut, and Metric Embeddings Date: 3/21/2007

Fréchet embeddings of negative type metrics

Lecture 18: March 15

Doubling metric spaces and embeddings. Assaf Naor

On metric characterizations of some classes of Banach spaces

Lecture: Expanders, in theory and in practice (2 of 2)

16 Embeddings of the Euclidean metric

Basic Properties of Metric and Normed Spaces

Euclidean Quotients of Finite Metric Spaces

On-line embeddings. Piotr Indyk Avner Magen Anastasios Sidiropoulos Anastasios Zouzias

An Improved Approximation Algorithm for Requirement Cut

Hardness of Embedding Metric Spaces of Equal Size

Algorithms, Geometry and Learning. Reading group Paris Syminelakis

On the Distortion of Embedding Perfect Binary Trees into Low-dimensional Euclidean Spaces

On Stochastic Decompositions of Metric Spaces

Similarity searching, or how to find your neighbors efficiently

Embeddings of finite metric spaces in Euclidean space: a probabilistic view

U.C. Berkeley CS294: Spectral Methods and Expanders Handout 11 Luca Trevisan February 29, 2016

Partitioning Algorithms that Combine Spectral and Flow Methods

MANOR MENDEL AND ASSAF NAOR

Multi-Embedding and Path Approximation of Metric Spaces

Bourgain s Theorem. Computational and Metric Geometry. Instructor: Yury Makarychev. d(s 1, s 2 ).

CMPUT 675: Approximation Algorithms Fall 2014

Lecture 10: Bourgain s Theorem

Notes taken by Costis Georgiou revised by Hamed Hatami

Small distortion and volume preserving embedding for Planar and Euclidian metrics Satish Rao

An Algorithmist s Toolkit Nov. 10, Lecture 17

Optimal compression of approximate Euclidean distances

Finite Presentations of Hyperbolic Groups

Computational Metric Embeddings. Anastasios Sidiropoulos

Integer Linear Programs

1 Dimension Reduction in Euclidean Space

Some Useful Background for Talk on the Fast Johnson-Lindenstrauss Transform

18.409: Topics in TCS: Embeddings of Finite Metric Spaces. Lecture 6

Ramsey partitions and proximity data structures

Introduction to Real Analysis Alternative Chapter 1

Ramsey partitions and proximity data structures

Markov chains in smooth Banach spaces and Gromov hyperbolic metric spaces

Ordinal Embedding: Approximation Algorithms and Dimensionality Reduction

Local Global Tradeoffs in Metric Embeddings

Metric structures in L 1 : Dimension, snowflakes, and average distortion

Partial cubes: structures, characterizations, and constructions

Metric spaces and metrizability

Math 320-2: Midterm 2 Practice Solutions Northwestern University, Winter 2015

Randomized Algorithms

MATH 4200 HW: PROBLEM SET FOUR: METRIC SPACES

Metric spaces admitting low-distortion embeddings into all n-dimensional Banach spaces

Economics 204 Fall 2011 Problem Set 1 Suggested Solutions

Immerse Metric Space Homework

Poorly embeddable metric spaces and Group Theory

A Union of Euclidean Spaces is Euclidean. Konstantin Makarychev, Northwestern Yury Makarychev, TTIC

Non-Asymptotic Theory of Random Matrices Lecture 4: Dimension Reduction Date: January 16, 2007

CSE 206A: Lattice Algorithms and Applications Spring Minkowski s theorem. Instructor: Daniele Micciancio

An algorithm to increase the node-connectivity of a digraph by one

SCALE-OBLIVIOUS METRIC FRAGMENTATION AND THE NONLINEAR DVORETZKY THEOREM

Cuts & Metrics: Multiway Cut

Spherical Euclidean Distance Embedding of a Graph

Lipschitz p-convex and q-concave maps

Using the Johnson-Lindenstrauss lemma in linear and integer programming

25 Minimum bandwidth: Approximation via volume respecting embeddings

Hierarchical Clustering via Spreading Metrics

MA651 Topology. Lecture 10. Metric Spaces.

Math 5210, Definitions and Theorems on Metric Spaces

Chapter 11. Approximation Algorithms. Slides by Kevin Wayne Pearson-Addison Wesley. All rights reserved.

Cographs; chordal graphs and tree decompositions

The Lefthanded Local Lemma characterizes chordal dependency graphs

Notes 6 : First and second moment methods

Metric decomposition, smooth measures, and clustering

Math 421, Homework #9 Solutions

Aditya Bhaskara CS 5968/6968, Lecture 1: Introduction and Review 12 January 2016

Euclidean distortion and the Sparsest Cut

Constructive bounds for a Ramsey-type problem

Packing and decomposition of graphs with trees

The Lovász Local Lemma : A constructive proof

arxiv: v1 [cs.cg] 15 Jul 2008

Speeding up the Dreyfus-Wagner Algorithm for minimum Steiner trees

MATH 51H Section 4. October 16, Recall what it means for a function between metric spaces to be continuous:

Chapter 2. Metric Spaces. 2.1 Metric Spaces

The following definition is fundamental.

Approximation Algorithms for Min-Sum k-clustering and Balanced k-median

ASYMPTOTIC DIOPHANTINE APPROXIMATION: THE MULTIPLICATIVE CASE

Lecture Notes in Advanced Calculus 1 (80315) Raz Kupferman Institute of Mathematics The Hebrew University

CSC Linear Programming and Combinatorial Optimization Lecture 10: Semidefinite Programming

FREE SPACES OVER SOME PROPER METRIC SPACES

Algorithms and Data Structures 2016 Week 5 solutions (Tues 9th - Fri 12th February)

ADJOINTS, ABSOLUTE VALUES AND POLAR DECOMPOSITIONS

Optimization and Optimal Control in Banach Spaces

- Well-characterized problems, min-max relations, approximate certificates. - LP problems in the standard form, primal and dual linear programs

Tree-width and algorithms

Coarse differentiation and multi-flows in planar graphs

Inderjit Dhillon The University of Texas at Austin

ANALYSIS WORKSHEET II: METRIC SPACES

Some Background Material

THEOREMS, ETC., FOR MATH 515

arxiv: v4 [cs.cg] 31 Mar 2018

Duality of multiparameter Hardy spaces H p on spaces of homogeneous type

ON TWO QUESTIONS ABOUT CIRCULAR CHOOSABILITY

Branchwidth of graphic matroids.

Duality of LPs and Applications

Transcription:

Finite Metric Spaces & Their Embeddings: Introduction and Basic Tools Manor Mendel, CMI, Caltech 1 Finite Metric Spaces Definition of (semi) metric. (M, ρ): M a (finite) set of points. ρ a distance function satisfying x M ρ(x, x) = 0, x, y M ρ(x, y) = ρ(y, x), x, y, z M ρ(x, z) ρ(x, y) + ρ(y, z). CS motivation. Finite metric spaces arise naturally in combinatorial objects, and algorithmic questions. For example, as the shortest path metrics on graphs. We will also see less obvious connections. Properties of finite metrics. The following properties have been investigated: Dimension, extendability of Lipschitz and Hölder functions, decomposability, Inequalities satisfied by the metric, short representations, additive distortion of embedding, (multiplicative) distortion of embeddings. We will focus on the last property. Embedding. A mapping f : (M, ρ) (H, ν) of a metric space M into a host metric space H, that (hopefully) preserves the geometry of M (usually distances). Distortion of embedding. The distortion of f : (M, ρ) (H, ν) is the least K 1 for which exists C > 0 such that x, y M, C ρ(x, y) ν(f(x), f(y)) K C ρ(x, y). (1) The distortion of f is denoted by dist(f). C is a scaling factor. Another way to define the distortion: The Lipschitz constant of a mapping f : (M, ρ) (H, ν) is Then dist(f) = f Lip f 1 Lip. f Lip = max x,y M ρ(x,y)>0 ν(f(x), f(y)). ρ(x, y) 1

Example of hosts spaces. l p is the set of real valued sequences (x i ) i N for which i x i p <. The l p norm is defined as x p = ( i x i p ) 1/p. Of particular interest are the cases p = 1, 2,. l 2 is called Hilbert space. Euclidean spaces are finite subspaces of Hilbert space. Non-embeddability example. Consider the (SP) metric of the graph C 4, a cycle on four nodes. Is there an isometric embedding of C 4 in Hilbert space? No: Denote the vertices on the C 4 by a 1,..., a 4. Suppose an isometric embedding exists. Note that ρ(a 1, a 3 ) = ρ(a 1, a 2 ) + ρ(a 2, a 3 ), hence the triangle inequality holds with equality, which means (for Euclidean spaces) that f(a 2 ) is in the middle of the segment [f(a 1 ), f(a 3 )]. Analogously, f(a 4 ) is in the middle of the segment [f(a 1 ), f(a 3 )]. Hence f(a 2 ) = f(a 4 ), a contradiction. [FIGURE] Interestingly, the natural embedding of C 4 as a square in the plain is the best embedding in Hilbert space, and its distortion is 2. Example of an algorithmic application. Assume we are given n-point set in X l d 1, and we would like to compute diam(x) = max x,y X x y 1. A straightforward algorithm is to compute the distance between all pairs in X and taking the maximum. This algorithm takes O(dn 2 ) time. We next show an improved algorithm when n d. We first exhibits an isometric embedding f : X l d, where d = 2 d. We then compute diam(f(x)). The second follows from max f(x) f(y) = max x,y X max x,y X j max j f(x) j f(y) j = max f(x) j f(y) j = max(max f(x) j min f(x) j). x,y X j x X x X Hence, in l d the diameter can be computed in O(nd ) time. The embedding f : l d 1 l {±1}d is defined as follows: (f(x 1,..., x d )) σ = j σ jx j, where for σ {±1} d. Computing this embedding can be clearly carried in O(ndd ) time (and actually also in O(nd ) time). Note that this embedding has additional nice properties: isometric, linear, and of the whole l d 1 space. In general, it will be rare to encounter embeddings with all these properties. 2 Bourgain Embedding As we have just seen, not all metrics are isometrically embedded in Hilbert space. In fact, it was shown that there exist n-point metrics for which the Euclidean distortion is Ω(log n) (for any n > 1). An important theorem of Bourgain states that this bound is tight. Thm. 1. Any n-point metric space can be embedded in l 2 with distortion O(log n). It has been shown that Bourgain theorem implies an efficient randomized algorithm for the embedding, and that the bound also holds for embedding in l p, p 1. 2

The sparest cut problem. An algorithmic application of Bourgain s theorem: Let G = (V, E) be a graph. A cut in G is a partition of V into two nonempty subsets A and B = V \A. The density of the cut (A, B) is e(a,b), where e(a, B) is the number of edges crossing the A B cut. Given G, we look for a cut of the smallest possible density. This problem is NP-hard, and we now present an efficient algorithm that produces a cut of density at most O(log V ) the density of the sparsest cut. We reformulate the problem using cut metrics. A cut metric τ A on V corresponds to the cut (A, V \ A) and is defined as τ A (u, v) = 1 when u and v are not on the same side of the cut (A, V \ A), and 0 otherwise. The sparsest cut problem asks to find (u,v) E τ(u, v) min, τ τ(u, v) u,v V where the minimum is taken over all cut metrics. We next relax the problem by allowing τ to be arbitrary metric. By scaling we obtain the following linear programming problem min τ((u, v)) (u,v) E Subject to u,v V τ(u, v) 1 τ satisfies the ineq. τ 0 This LP can be solved in polynomial time. The value of the optimal solution is at most the sparsest cut, but it is not necessarily a cut. We round it by embedding the resulting metric τ 0 into l m 1 (m = O(log 2 n)) using Bourgain s theorem. We obtain a metric τ 1 in l m 1,satisfying τ 1 τ 0 O(log V )τ 1 An l 1 metric metric can be written as a positive sum of cut metrics (i.e., it is in the cut-cone). Thus, τ 1 = α i θ i. i α i > 0, and θ i is a cut metric (such a decomposition is easily computed). Note that O(log V ) e E τ 0(e) u,v V τ 0(u, v) e E τ 1(e) u,v V τ 1(u, v) = i α i (e E θ i(e) ) i α i ( u,v V θ i(u, v) ) min i e E θ i(e) u,v V θ i(u, v). The RHS is easily computed and its value is at most O(log V ) the value of the LP (which is at most the density of the sparsest cut. 3

3 Embedding in probabilistic trees Tree metrics. A tree metric is a metric given by the SP on a weighted tree. A finite ultrametric is a metric induced on the leaves of a rooted tree T, where each node v of the tree has label (v). The distance between two leaves x, y is defined by (lca(x, y)). [Figure] Ultrametric is a special type of tree metric. Since tree metrics, and in particular ultrametrics are relatively simple, it s usually simpler to devise algorithms to these classes then to general metrics. Unfortunately simple metrics (like cycles) do not embed well in tree metrics. Embedding into probabilistic trees. A convex combination of metrics is also a metric. Efficient embedding into a convex combination allows for a randomized algorithm to sample a tree metric from the distribution and apply the tree algorithm. However, for cost approximation algorithms, plain distortion is usually not enough. An embedding f : (X, ρ) i α iν i into convex combination of metrics ν i is called probabilistic embedding if ρ ν i. The distortion of the embedding is the minimum β 1 such that i α iν i β ρ. Thm. 2. Any n point metric can be probabilistically embedded in a convex combination of ultrametrics, with distortion O(log n). Algorithmic Example: Metric k-median. Algorithmic clustering problem: Given n- point metric space (M, ρ), and an integer k, find k centers C = {c 1,..., c k } M that minimizes the expression min x Mρ(x, C), C M C k where ρ(x, C) = min y C ρ(x, y). A natural clustering problem. An algorithm for ultrametrics. We may assume that the ultrametric tree is binary (by splitting a vertex with more than two children into a chain with the same labels) Apply a Dynamic Programming algorithm that computes for each vertex v, k {0,..., n}, compute COST(v, k), which is the optimal cost for serving the vertices in the subtree rooted at v using k centers, The algorithm works bottom-up. Let v be vertex. Denote by l(v) the number of leaves in the subtree rooted at v. If v is a leaf then COST(v, 0) =, and COST(v, k) = 0 for k 1. If v has two children v 1, v 2, then { COST(v, k) = min min k 1,k 2 1 k 1 +k 2 =k A randomized alg for the k-median COST(v 1, k 1 ) + COST(v 2, k 2 ), } COST(v 1, k) + l(v 2 ) (v), COST(v 2, k) + l(v 1 ) (v). Med(M, k). Sample an ultrametric T from the embedding of M into probabilistic trees. 4

Find an r approximation of the k-median problem on trees. Output this solution. We claim that this is O(log n)r approximation of the optimal solution. To see this, let O M be the optimal solution for M, and let A T the solution of the approximation alg for T. Then T cost M (A T ) cost T (A T ), since the distance in T dominates the distances in M. Also, cost T (A T ) r cost T (O M ), since A is an r-approximation to the optimal cost in T (which is at most cost T (O)). Since O doesn t depend on the sampling of the trees, we have by linearity of expectation, Hence E T [cost T (O M )] O(log n)cost M (O M ). E T [cost M (A T )] O(r log n)cost M (O M ). and this is what we have looked for, since the LHS is the expected cost of the randomized algorithm. Remarks. 1. This method works for many cost minimization problems on metrics in which the cost is conically depends on the distances. 2. This algorithm was the first non-trivial approximation algorithm for the k-median in general finite metrics. By now, a constant approximation algorithm is known. 4 The Johnson-Lindenstrauss Lemma. Thm. 3. For any n-point in l d 2, there exists a linear transformation F : l d 2 l 16ε 2 log n 2. Even more, there exists a probability distribution D of linear transformations F : l d 2 l 16ε 2 log n 2, such that for any n point set X l d 2, [ dist(f X : X l d 2) 1 + ε ] 0.9 Pr F D Algorithmic application. Consider computing the diameter of n points in l d 2, where d is large (say, d = n 1). The straightforward algorithm takes O(dn 2 ) time. By applying dimension reduction first, we obtain an algorithm that computes 1 + ε approximation to the diameter, and the running time is O(nd log n + n 2 log n). Remark. Better algorithms are known, with running time O(n 2 ε + dn). 5

5 A proof of a variant of Bourgain embedding We will prove Thm. 4. For any b N, any n-point metric (M, ρ) can be embedded into l d with distortion c = 2b 1, for d = O(bn 1/b log n) It implies O(log n) distortion embedding in l O(log2 n, which implies O(log 2 n) embedding in l 2. The proof is somewhat simpler, but the technique is the same. We start with an isometric embedding f : M l n. We index the coordinates of l n using the points of M, and define f(y) x = ρ(x, y). Then No shrinkage of distances: f(x) f(y) = max z M ρ(x, z) ρ(y, z) ρ(x, y) ρ(y, y) = ρ(x, y). No expansion of distances: By the triangle inequality ρ(z, y) ρ(z, x) ρ(x, y), so f(x) f(y) = max z M ρ(x, z) ρ(y, z) ρ(x, y). In the general case, the coordinates will be indexed by subsets of M, and f A (y) = ρ(a, y). The no expansion argument (using the triangle inequality) still holds, but the no shrinkage property will hold only approximately. The idea: Find (randomly) subsets A A such that for every pair x, y M, there exists a radius r and a set A A for which B(x, r) A, whereas B(y, r + ρ(x, y)/c) A = [FIGURE]. In this case ρ(a, x) ρ(a, y) ρ(x, y)/c, as needed. Construction of A. We set p j = n j/b. For each j {1,..., b}, we construct 10 log nn 1/b sets, each by sampling points independently with probability p j (this gives O(bn 1/b log n) coordinates). Analysis. Fix a pair of points x, y M. Denote by b i (x) = B(x, i d(x, y)/c), and b i (y) = B(y, i d(x, y)/c). By the pigeon-hole principle, there must be i 0, j 0 for which or b i0 (y) n j 0/b, and b i0 +1(x) n (j 0+1)/b, b i0 (x) n j 0/b, and b i0 +1(y) n (j 0+1)/b, (say the first case happened). Consider a set A chosen with probability p j0 +1. Pr[A b i0 +1(x) = ] (1 p j0 +1) n(j 0 +1)/b e 1. Pr[A b i0 (y) ] 1 (1 p j0 +1) nj 0 /b 1 e n 1/b n 1/b. Since b i0 +1(x) and b i0 (y) are pairwise disjoint, the two events are independent, and therefore they happen simultaneously with probability Θ(n 1/b ). Since we repeat this experiment O(n 1/b log n) times, the probability that one of the sets A A will satisfies both events is 1 0.1 n 2. Hence with probability 0.9, all pairs has a set A A which satisfies both events, and this set catches 1/c fraction of the distance between the pair, as needed. 6

References [1] A. Gupta, and R. Ravi. Algorithmic Applications of Metric Embeddings. A course given in CMU, 2003. [2] P. Indyk. Algorithmic Aspects of Geometric Embeddings FOCS 2001 [3] P. Indyk, J. Matoušek. Low distortion embedding of finite metric spaces. Chapter 8 in the Handbook of Discrete and Computational Geometry, second edition, 2004. [4] J. Matoušek. Lectures on Discrete Geometry Chapter 15. 7