A Randomized Approximate Nearest Neighbors Algorithm - a short version

Size: px
Start display at page:

Download "A Randomized Approximate Nearest Neighbors Algorithm - a short version"

Transcription

1 We present a ranomize algorithm for the approximate nearest neighbor problem in - imensional Eucliean space. Given N points {x } in R, the algorithm attempts to fin k nearest neighbors for each of x, where k is a user-specifie integer parameter. The algorithm is iterative, an its CPU time requirements are proportional to T N ( (log )+ k ( + log k) (log N)) + N k 2 ( + log k), with T the number of iterations performe. The memory requirements of the proceure are of the orer N ( + k). A byprouct of the scheme is a ata structure, permitting a rapi search for the k nearest neighbors among {x } for an arbitrary point x R. The cost of each such query is proportional to T ( (log ) + log(n/k) k ( + log k)), an the memory requirements for the requisite ata structure are of the orer N ( + k) + T ( + N). The algorithm utilizes ranom rotations an a basic ivie-an-conquer scheme, followe by a local graph search. We analyze the scheme s behavior for certain types of istributions of {x }, an illustrate its performance via several numerical examples. A Ranomize Approximate Nearest Neighbors Algorithm - a short version Peter W. Jones, Anrei Osipov, Vlaimir Rokhlin Research Report YALEU/DCS/TR-1439 Yale University January 13, 2011 This author s research was supporte in part by the DMS grant # an the ONR grants #N , #N ; this author s research was supporte in part by the AFOSR grant #FA ; this author s research was supporte in part by the ONR grant #N an the AFOSR grant #FA Approve for public release: istribution is unlimite. Keywors: Approximate nearest neighbors, ranomize algorithms, fast ranom rotations 1

2 1 Introuction In this paper, we escribe an algorithm for fining approximate nearest neighbors (ANN) in -imensional Eucliean space for each of N user-specifie points {x }. For each point x, the scheme prouces a list of k suspects, that have high probability of being the k closest points (nearest neighbors) in the Eucliean metric. Those of the suspects that are not among the true nearest neighbors, are close to being so. We present several measures of performance (in terms of statistics of the k chosen suspecte nearest neighbors), for ifferent types of ranomly generate ata sets consisting of N points in R. Unlike other ANN algorithms that have been recently propose (see e.g. [1]), the metho of this paper oes not use locality-sensitive hashing. Instea we use a simple ranomize ivie-an-conquer approach. The basic algorithm is iterate several times, an then followe by a local graph search. The performance of any fast ANN algorithm must eteriorate as the imension increases. While the running time of our algorithm only grows as log, the statistics of the selecte approximate nearest neighbors eteriorate as the imension increases. We provie bouns for this eterioration (both analytically an empirically), which occurs reasonably slowly as increases. While the actual estimates are fairly complicate, it is reasonable to say that in 20 imensions the scheme performs extremely well, an the performance oes not seriously eteriorate until is approximately 60. At = 100, the egraation of the statistics isplaye by the algorithm is quite noticeable. An outline of our algorithm is as follows: 1. Choose a ranom rotation, acting on R, an rotate the N given points. 2. Take the first coorinate, an ivie the ata set into two boxes, where the boxes are ivie by fining the meian in the first coorinate. 3. On each box from Step 2, we repeat the subivision on the secon coorinate, obtaining four boxes in total. 4. We repeat this on coorinates 3, 4, etc., until each of the boxes has approximately k points. 5. We o a local search on the tree of boxes to obtain approximately k suspects, for each point x. 6. The above proceure is iterate T times, an for each point x, we select from the T k suspects the k closest iscovere points for x. 7. Perform a local graph search on the collections of suspects, obtaine in Step 6 (we call this local graph search supercharging ). Among k 2 caniates obtaine from the local graph search, we select the best k points an eclare these the suspecte approximate nearest neighbors, or suspects. The ata structure generate by this algorithm allows one to fin, for a new ata point y, the k suspecte approximate nearest neighbors in the original ataset. This search is quite rapi, as we nee only follow the alreay generate tree structure of the boxes, obtaine in 2

3 the steps liste above. One can easily see that the epth of the binary tree, generate by Steps 1 through 4, is log 2 (N/k). This means that we can use the T trees generate, an then pass to Step 7. Almost all known techniques for solving ANN problems use tree structures (see e.g. [1], [2], [3]). Two apparently novel features of our metho are the use of fast ranom rotations (Step 1), an the local graph search (Step 7), which ramatically increases the accuracy of the scheme. We use the Fast Fourier Transform to generate our ranom rotations, an this accounts for the factor of log that appears in the running time. Our use of ranom rotations replaces the usual proection argument use in other ANN algorithms, where one proects the ata on a ranom subspace. As far as we know, the use of fast rotations for applications of this type appears first in [3] (see [4] an the references therein for a brief history). The use of ranom rotations (as in our paper) or ranom proections (as use elsewhere in ANN algorithms) takes avantage of the same unerlying phenomenon; namely the Johnson-Linenstrauss Lemma. (The JL Lemma roughly states that proection of N points on a ranom subspace of imension C(ε) (log N) has expecte istortion 1 + ε, see e.g. [5].) We have chosen to use ranom rotations in place of the usual ranom proections generate by selecting ranom Gaussian vectors. The fast ranom rotations require O( (log )) operations, which is an improvement over methos using ranom proections (see [6], [7]). The N k lookup table arising in Step 7 is the aacency matrix of a graph whose vertices are the points {x }. In Step 7 we perform a epth one search on this graph, an obtain k + k 2 caniates (of whom we select the suspects ). This accounts for the factor of k 2 in the running time. Due to egraation of the running time, we have chosen not to perform searches of epth greater than one. The algorithm has been teste on a number of artificially generate point istributions. Results of some of those tests are presente below. The paper is organize as follows. In the first section, we summarize the mathematical an numerical facts to be use in subsequent sections. In the secon section, we escribe the Ranomize Approximate Nearest Neighbors algorithm (RANN) an analyze its cost an performance. In the thir section, we illustrate the performance of the algorithm with several numerical examples. 2 Mathematical Preliminaries In this section, we introuce notation an summarize several facts to be use in the rest of the paper. 2.1 Degree of contact Suppose that L > 0 are positive integers. Suppose further that σ = σ 1... σ L, µ = µ 1...µ L, σ i, µ {+, } (1) are two wors of symbols +, of length L. We efine the egree of contact Con(σ, µ) between σ an µ to be the number of positions at which the corresponing symbols are 3

4 ifferent. In other wors, Con(σ, µ) = {i : 1 i L, σ i µ i }. (2) In a mil abuse of notation, we say that two isoint sets Aσ an Aµ (or their elements) have egree of contact if Con(σ, µ) =. For example, x an y have egree of contact 1 if x Aσ, y Aµ an σ, µ iffer at precisely one symbol. 2.2 Pseuoranom orthogonal transformations In this section, we escribe a fast metho (presente in [6], [7]) for the generation of ranom orthogonal transformations an their application to arbitrary vectors. Suppose that, M 1, M 2 > 0 are positive integers. We efine a pseuoranom - imensional orthogonal transformation Θ as a composition of M 1 +M 2 +1 linear operators M 1 Θ = =1 P () Q () P () F () M 1 +M 2 =M 1 +1 Q () P (). (3) The linear operators P () : R R with = 1,...,M 1 +M 2 are efine in the following manner. We generate permutations π 1,...,π M1 +M 2 of the numbers {1,...,}, uniformly at ranom an inepenent of each other. Then for all x R, we efine P () x by the formula ( ) x (i) = x(π (i)), i = 1,...,. (4) In other wors, P () permutes the coorinates of the vector x accoring to π. P () can be represente by a matrix P, efine by the formula { 1 l = π (k), P (k, l) = (5) 0 l π (k), for k, l = 1,...,. Obviously, the operators P () are orthogonal. The linear operators Q () : R R with = 1,...,M 1 +M 2 are efine as follows. We construct ( 1) (M 1 + M 2 ) inepenent pseuoranom numbers, θ (1),...,θ ( 1) with = 1,...,M 1 + M 2, uniformly istribute in (0, 2π). Then we efine the auxiliary linear operator Q,k : R R for k = 1,..., 1 to be the planar rotation of the coorinates k an k + 1 by the angle θ (k). In other wors, for all x R ( ) k (Q,k (x)) = k + 1 ( ) ( ) cos(θ (k)) sin(θ (k)) x(k), (6) sin(θ (k)) cos(θ (k)) x(k + 1) an the rest of the coorinates of Q,k (x) coincie with those of x. We efine Q () formula by the Q () = Q, 1 Q, 2 Q,1. (7) 4

5 Obviously, the operators Q () are orthogonal. The linear operator F () : R R is efine as follows. First suppose that is even an that 2 = /2. We efine the 2 2 iscrete Fourier transform matrix T by the formula T(k, l) = 1 2 exp [ 2πi(k 1)(l 1) 2 ], (8) where k, l = 1,..., 2 an i = 1. The matrix T represents a unitary operator C 2 C 2. We then efine the one-to-one linear operator Z : R C 2 by the formula x(1) + i x(2), Zx = x(3) + i x(4), (9) x(2 2 1) + i x(2 2 ) for all x R. Eventually, we efine F () by the formula F () = Z 1 T Z (10) for even. If is o, we efine F () x for all x R by applying F ( 1) to the first 1 coorinates of x an leaving its last coorinate unchange. Obviously, the operators T, Z, efine by (8), (9), respectively, preserve the norm of any vector x R. Therefore, F () is a real orthogonal transformation R R. The cost of the generation of a ranom permutation (see e.g. [8]) is O() operations. The cost of the application of each P () to a vector x R is obviously operations ue to (4). The cost of generation of 1 uniform ranom variables is O() operations. Also, the cost of application of each Q () to a vector x R is O() operations ue to (7). Finally, the cost of the fast iscrete Fourier transform is O( log ) operations, an the cost of the application of F () to a vector x R is O( log ) operations ue to (8), (9) an (10). Thus the cost of the generation of Θ efine via (3) is Cost(Θ) = O ( (M 1 + M 2 + log )). (11) Moreover, the cost of application of Θ to a vector x R is also given by the formula (11). Remark 1. It was observe that if M 1 an M 2 in (3) are chosen such that M 1 +M 2 log, then the istribution of Θ is close to the uniform istribution on the group O(, R) of orthogonal transformations from R to R. Remark 2. The use of the Haamar matrix (without 2 2 rotations) appears in a relate problem stuie by Ailon an Liberty [9]. 3 Ranomize Approximate Nearest Neighbors algorithm In this section, we escribe the Nearest Neighbor Problem an present a fast ranomize algorithm for its solution. 5

6 3.1 The Nearest Neighbor Problem Suppose that an k < N are positive integers an suppose that B = {x 1, x 2,...,x N } R (12) is a collection of N points in R. We are intereste in fining the k nearest neighbors of each point x i. The istance between the points is the stanar Eucliean istance. For each x i, one can compute in a straightforwar manner the istances to the rest of the points an thus fin the nearest neighbors. However, the total cost of the evaluation of the istances alone is O( N 2 ), which makes this naive approach prohibitively expensive when N is large. We propose a faster approximate algorithm for the solution of this problem. 3.2 Informal escription of the algorithm Initial selection The key iea of our algorithm is the following simple (an well known) observation. Suppose that for each x i we have foun a small subset V i of B such that a point insie V i is more likely to be among the k nearest neighbors of x i than a point outsie V i. Then it is reasonable to look for the nearest neighbors of each x i only insie V i an not among all the points. The nearest neighbors of x i in V i, which can be foun by irect scanning, are referre to as its suspecte approximate nearest neighbors, or suspects, as oppose to the true nearest neighbors { x t(i,) }. Of course, many of the k true nearest neighbors of x i might not be among its suspects. However, one can re-select V i to obtain another list of k suspects of x i. The initial guess is improve by taking the best k points out of the two lists. This scheme is iterate to successively improve the list of suspects of each x i. The performance of the resulting iterative ranomize algorithm amits the following crue analysis. Suppose that the size of V i is α N, with α 1. Suppose also that the number of the true nearest neighbors of x i insie V i is roughly β k, with α < β < 1. If the choice of V i is fairly ranom, then orer O(1/β) iterations of the algorithm are require to fin most of the true nearest neighbors of each x i. Temporarily neglecting the cost of the construction of V i, this results in O ( (α/β) N 2) operations instea of O ( N 2) operations for the naive algorithm. If α β, the improvement can be substantial. Our construction of V i s is base on geometric consierations. First, we shift all of the points to place their center of mass at the origin an apply a ranom orthogonal linear transformation on the resulting collection. Then, we choose a real number y(1) such that the first coorinate of half of the points in B is less than y(1). In other wors, We ivie all the points into two isoint sets {x B : x(1) < y(1)} = N/2. (13) B = {x B : x(1) < y(1)}, B + = {x B : x(1) y(1)}. (14) 6

7 Obviously, the sizes of B an B + are the same if N is even an iffer by one if N is o. Next, we set y + (2) to be a real number such that the secon coorinate of half of the points in B + is less than y + (2), i.e. {x B + : x(2) < y + (2)} = N/4. (15) We split B + into two isoint sets B + an B ++ by the same principle, e.g. B + = {x B + : x(2) < y + (2)}, B ++ = {x B + : x(2) y + (2)}. (16) We construct B an B + in a similar fashion by using a real number y (2) such that the secon coorinate of half of the points in B is less than y (2), i.e. {x B : x(2) < y (2)} = N/4. (17) Each of the four boxes B, B +, B +, B ++ contains either N/4 or N/4 + 1 points. Then we repeat the subivision by splitting each of the four boxes into two by using the thir coorinate, an so on. We procee until we en up with a collection of 2 L boxes {Bσ} with k or k + 1 points in each box. Here the box inex σ is a wor of symbols +, of length L as in (1) an L is a positive integer such that k 2 L N < k 2 L+1. In other wors, L is efine via the inequality k 2 L N < k 2 L+1. (18) Obviously, the sets { Bµ } constitute a complete binary tree of length L, whose noes are inexe by wors µ of symbols +, of length up to L. The set B is at the root of this tree, the sets B an B + are at the secon level, an so on. At the last level of the tree, there are 2 L boxes. The epth of the tree equals to L + 1. The notion of egree of contact (2) extens to the collection {Bσ} of boxes. Suppose that x i is in Bσ. Obviously, the higher egree of contact of two boxes Bσ an Bµ is, the less likely a point of Bµ will be among the k nearest neighbors of x i. Motivate by this observation, we efine the set V i as V i = { x Bµ : Con(σ, µ) 1 }. (19) In other wors, V i is the union of the box Bσ containing x i an L boxes whose egree of contact with Bσ is one. Thus for each i = 1,...,N, the set V i contains about k (L + 1) points Supercharging In the previous subsections, we have escribe an iterative scheme for the selection of suspects (suspecte approximate nearest neighbors) for each of the points x i in B. Suppose now that after T iterations of this scheme, the list x s(i,1),...,x s(1,k) of k suspects of each point x i has been generate. This list can be improve by a proceure we call supercharging. The iea of supercharging is base on the following observation. A true nearest neighbor of x i, misse by the scheme escribe above, might be among the suspects of one of x s(i,1),...,x s(1,k). This leas to the following obvious proceure. 7

8 For each x i, we enote by A i the list of suspects of all x s(i,1),...,x s(i,k). A i contains k 2 points, with possible repetitions. We compute the square of the istances from x i to each point in A i an fin the k nearest neighbors x t(i,1,ai ),...,x t(i,k,ai ) of x i in A i. Then we eclare the (upate) suspects of x i to be the best k points out of the two lists { } k x s(i,) =1 an { k x t(i,,ai )} =1. In other wors, supercharging is a epth one search on the irecte graph, whose vertices are the points {x i } an whose N k aacency matrix is the suspects inices {s(i, )}, with i = 1,...,N an = 1,...,k Overview We conclue this section with a list of the principal steps of the algorithm. Given the collection {x i } N i=1 of points in R, we perform the following operations. 1. Subtract from each x i the center of mass of the collection. 2. Choose a ranom orthogonal linear transformation Θ an set x i = Θ(x i ) for all i = 1,...,N. 3. Construct 2 L boxes {Bσ} as escribe above. 4. For each x i efine the set V i via (19). 5. Upate the suspects x s(i,1),...,x s(i,k) of x i by using its true nearest neighbors in V i. 6. Steps 2-5 are repeate T times. 7. For each x i, perform supercharging Query for a new point Suppose that we are given a new point y R, an we nee to fin its k nearest neighbors in B = {x 1,...,x N }. In this subsection, we escribe a rapi proceure to fin k approximate nearest neighbors of y. This proceure uses the following information, available on the th iteration of the algorithm, for = 1,...,T: 1. The orthogonal linear transformation Θ (), generate on the th iteration of the algorithm. 2. The collection of boxes { B σ () }, generate on the th iteration of the algorithm. To fin k approximate nearest neighbors of the new point y among the points of B, we perform the following operations. First, we apply Θ (1) on y, where Θ (1) is the orthogonal linear transformation of the first iteration of the algorithm. The resulting vector is enote by y (1), in other wors, y (1) = Θ (1) (y). (20) 8

9 { Next, in the collection of boxes B (1) σ }, generate on the first iteration of the algorithm, we fin the box B (1) σ(1) that has egree of contact zero with y(1). Note that if y ha belonge to B in the first place, then on the first iteration of the algorithm y (1) woul have belonge to B (1) σ(1). The box B (1) σ(1) has k or k + 1 points. Suppose that V (1) σ(1) having egree of contact zero or one with B (1) σ(1) (L + 1) k points, where L is efine via (18). is is the union of the boxes (see (19)). Recall that V (1) σ(1) has about We efine the set V (2) σ(2) in a similar manner, by using the ata of the secon iteration of the algorithm. We apply the orthogonal transformation Θ (2) of the secon iteration on y (1) to obtain y (2), i.e. ( y (2) = Θ (2) y (1)) ( ) = Θ (2) Θ (1) (y), (21) { ue to (20). In the boxes } σ B (2) of the secon iteration, we fin the box B (2) σ(2), having egree of contact zero with y (2). Similar to V (1) (2) σ(1), the set V σ(2) has about (L+1) k points. We repeat this proceure to construct the sets V () σ() for = 3, 4,...,T, where T is the number of the iterations of the algorithm. Each V () σ() contains roughly (L + 1) k points. Finally, we efine the set A to be the union of all the sets V () σ(), in other wors, A = T =1 V () σ(). (22) The set A contains about T k (L + 1) points. The k nearest neighbors of y insie A are eclare to be the approximate nearest neighbors of y insie B. We note that to construct A we nee to store the corresponing ata on each iteration of the algorithm. Once the k suspects of y in B have been foun, they can be improve by performing supercharging on y only. 3.3 Cost analysis In this subsection, we analyze the cost of the algorithm in terms of number of operations. Also, we analyze the memory requirements of the algorithm. We recall that x 1,...,x N is a collection of N points in R an N k 2 L. 1. Centralizing the points costs O(N ). 2. Generating a pseuoranom orthogonal transformation an applying it to N points costs O(N (log )) (see (11)). 3. Constructing the binary tree of boxes of epth L = log 2 (N/k) costs O(N L). 4. The suspects of each point x i are selecte by computing the istances from x i to k (L + 1) points in V i (see (19)) an fining k minimal istances. Thus the cost of this step is O(N L k ( + log k)). 9

10 5. In supercharging, k 2 points are scanne to upate the suspects of each point. Thus the cost of supercharging is O(N k 2 ( + log k)). We conclue that the total cost of the algorithm is C RANN = O (T N ( (log ) + k ( + log k) (log N))) + O(N k 2 ( + log k)), (23) where T is the number of iterations. We observe that for fixe imension an number of require nearest neighbors k, the cost is O(T N log N), as oppose to O(N 2 ) of the naive approach. Also, the cost of supercharging is quaratic in the number of nearest neighbors for fixe imension an number of points N, which makes supercharging expensive relative to a single iteration of the principal part of the algorithm even for moerate k. Query for a new point y consists of performing all the steps of the algorithm on one point only. Consequently, ue to (23) the total cost of query for a new point is C query = O (T ( (log ) + k ( + log k) (log N))) + O(k 2 ( + log k)). (24) To etermine the memory requirements of the algorithm, we observe that to store N points in R we nee O(N ) memory, to store the inices of k nearest neighbors of each of N points we nee O(N k) memory, an to store a tree of boxes (with the corresponing orthogonal transformation) we nee O(N) memory. We must istinguish between two cases. In the first case, given N points {x i } in R, we are intereste in fining k nearest neighbors for each x i only. In other wors, no query for a new point will ever be requeste. Then, the tree of boxes can be estroye after each iteration. Hence, in this case the algorithm uses O(N ( + k)) memory. In other wors, in this case the memory requirements are minimal, in the sense that most of the memory is spent on the storage of input an output of the algorithm only. In the secon case, we know in avance that queries for new points will be requeste. Therefore we nee to store T trees of boxes an the corresponing orthogonal transformations. Thus, when queries for new points are allowe, the total memory requirements are of the orer M query = O (N ( + k + T)). (25) Remark 3. The factor log k in (23), (24) can be omitte, if the suspects of every x i are not require to be sorte accoring to their istance to x i. 3.4 Performance analysis In this subsection, we iscuss the performance analysis of the Ranomize Approximate Nearest Neighbor algorithm in the case of stanar Gaussian istribution of the points. To be more specific, we consier the collection of N inepenent stanar Gaussian - imensional ranom vectors x 1,...,x N, where the number of points is given by the formula N = k 2 L (26) 10

11 for some positive integer L > 0. We recall that for each point x i, the algorithm approximates the k true nearest neighbors { } k x t(i,) =1 of x i by k suspects { } k x s(i,). In orer to analyze =1 the quality of this approximation, we introuce a number of statistical quantities. First, we efine the average square of the istance from x i to its k true nearest neighbors by the formula D true i = 1 k k x i x t(i,) 2. (27) =1 Next, we efine the average square of the istance from x i to its k suspects by the formula D susp i = 1 k k x i x s(i,) 2. (28) =1 Finally, we efine the proportion of the true nearest neighbors of x i among its suspects by the formula prop i = 1 { } k x k t(i,) =1 { } k x s(i,). (29) =1 We analyze a single iteration of the algorithm (selection of suspects without supercharging) by evaluating the expecte values of (27), (28), (29) numerically (see Theorems 1, 2, 3 below). On the other han, the performance of the algorithm can be stuie empirically, by running the algorithm on artificially generate sets of points. In the rest of this subsection, we outline the results of the analysis (see [10] for etails). We start with introucing some notation. Suppose that > 0 is a positive integer, an λ, α, β > 0 are positive real numbers. The istributions χ 2, χ2 (, λ) an B(α, β) are efine by their probability ensity functions, given respectively via the formulas f χ 2(t) = t/2 1 e t/2 2 /2, Γ(/2) t > 0, (30) f χ 2 (,λ)(t) = e λ/2 (λ/2) f! χ 2 (t), +2 t > 0, (31) f B(α,β) (t) = =0 Γ(α + β) Γ(α)Γ(β) tα 1 (1 t) β 1, 0 < t < 1. (32) Suppose that a > 0 is a positive real number. Suppose also that L > 0 are positive integers, an a R is a vector all of whose coorinates are positive. We efine the functions 11

12 h a, h + a, ha : R C via the formulas [ ia h 2 x a (x) = exp 1 2ix [ ia h + 2 x a (x) = exp 1 2ix ] [ ] erfc iax 2 1 2ix ] ha(x) = exp ix 1 2ix 1 1 2ix, (33) 2 h a (x), (34) 1 2ix ( ) a() 2 1 L 1 2ix =L+1 L L h + a() (x) + L h a(i) (x) h + a() (x). (35) =1 i=1 =1 i Also, we efine the function Ga : (0, ) R via the formula Ga(y) = 1 2π (L + 1) y 0 e ixt ha(x) x t. (36) The following theorem provies an analytical formula for the expectation E [ DN true ] (see (27)). Theorem 1. Suppose that, k, N > 0 are positive integers. Suppose further that DN true is efine by (27). Then its expectation is given by the formula E [ DN true ] = ( 1 k ) 1 F 1 k χ 2 (,λ) (t) f B(i,N i)(t) t f χ 2(λ) λ, (37) 0 i=1 0 where the functions f χ 2, f B(i,N i) are efine respectively by (30), (32), an F 1 is the χ 2 (,λ) inverse of the cf of χ 2 (, λ) (see (31)). The following theorem provies an approximation to the expectation E [ D susp ] N (see (28)). The error of this approximation is verifie via numerical experiments (see [10] for more etails). Theorem 2. Suppose that k > 0 an L > 0 are positive integers. We efine the real number D susp appr by the formula D susp appr = 0 Avg a S + (λ) [ 1 k k i=1 ( G 1 a i k L + k + 1 ) ] f χ 2(λ) λ, (38) where the function f χ 2 is efine via (30), the function G 1 a is the inverse of G a efine via (36), the set S + (λ) is the positive part of the -imensional hypersphere of raius λ, efine by the formula } S + {x (λ) = R : x 2 = λ, x() > 0, 1, (39) 12

13 an the average (Avg) is taken with respect to the ( 1)-imensional area. Then, E [ D susp ] ( N = D susp appr + O k 1/2), k. (40) In other wors, (40) hols, if we fix, L an let k. The following theorem provies an approximation to the expectation E[prop N ] (see (29)). The error of this approximation is verifie via numerical experiments (see [10] for more etails). Theorem 3. Suppose that k > 0 an L > 0 are positive integers. We efine the real number P appr by the formula P appr = (L + 1) 0 Avg a S + (λ) [ ( )] Ga F 1 χ 2 (,λ) (2 L ) f χ 2(λ) λ, (41) where the function f χ 2 is efine via (30), the function F 1 is the inverse of the cf of χ 2 (,λ) χ 2 (, λ), efine via (31), the function Ga is efine via (36), the set S + (λ) is efine via (39), an the average (Avg) is taken with respect to the ( 1)-imensional area. Then, ( E[prop N ] = P appr + O k 1/2), (42) where the real ranom variable prop N is efine via (29). In other wors, (42) hols, if we fix, L an let k. 4 Numerical Results This section has two principal purposes. First, we numerically evaluate the expectations of (27), (28), (29) by using Theorems 1, 2, 3 above. Secon, we emonstrate the performance of the algorithm empirically, by running it on sets of points, generate accoring to the Gaussian istribution. The choice of uniform or Hamming istributions instea of Gaussian results in very similar performance (see [10] for results an etails). The algorithm has been implemente in FORTRAN (Lahey 95 Linux version). The numerical experiments have been carrie out on a Lenovo laptop computer, with DualCore CPU 2.53 GHz an 2.9GB RAM. Experiment 1. In this experiment, we choose k = 30, an for L = 10, 15, 20, 25 set N via (26). Then, for ifferent values of between L an 110, we approximate the expectations E [ DN true ] [, E D susp] N, E[propN ] via the numerical evaluation of (37), (38), (41), respectively. The approximations are enote by Dnum, true Dnum, susp P num, respectively, an are accurate to roughly two ecimal igits. Also, we compute the ratio Ratio susp true = Dsusp num Dnum true. (43) 13

14 The results of Experiment 1 are shown in Figure 1. In Figure 1 (left), we plot Ratio susp true as a function of the imensionality, for each value of L. In Figure 1 (right), we plot P num as a function of the imensionality, for each value of L (the number of points N is etermine via (26)). This quantity estimates the average proportion of the suspects among the true nearest neighbors (foun by one iteration of RANN without supercharging). Several observations can be mae from Experiment 1 (see Figure 1) an from more etaile experiments by the authors. 1. The ratio Ratio susp true, efine via (43), slowly increases with the number of points N, for fixe values of number k of nearest neighbors an imensionality (see Figure 1, left). In other wors, the performance of RANN eteriorates with the number of points, if terms of Ratio susp true. However, as number of points is multiplie by 2, this ratio grows by a factor less than 1.1 for most values of on Figure 1 (left). 2. The ratio Ratio susp true actually ecreases with imensionality, for fixe k an N (see Figure 1, left). For example, for N 10 6, this ratio ecays from the value of about 1.57 at = 20 to roughly 1.3 at = 110. This observation is not surprising, since the number of points within the same fixe relative istance (e.g. twice the istance to true nearest neighbors) grows with the imensionality. 3. The proportion of true nearest neighbors among suspects ecays with for fixe k, N, as expecte (see Figure 1, right). However, even for = 40 an N 10 6, RANN correctly fins about 2.7% of true nearest neighbors, on merely one iteration without supercharging. In other wors, after 50 iterations without supercharging RANN will correctly etect about 75% = ( ) 50 (44) true nearest neighbors, for = 40 an N 10 6 (see also Experiment 2 below). Experiment 2. In this experiment, we run RANN with various parameters on ifferent sets of points an report on the resulting statistics. We choose the imensionality an the number of points N. Then we choose the number of nearest neighbors k, the number of iterations of the algorithm T > 0, an whether to perform the supercharging or not. Next, we generate N i.i.. stanar Gaussian vectors x 1,...,x N in R. We run RANN on x 1,...,x N. For each x i, RANN fins its k suspects, x s(i,1),...,x s(i,k). Also, for all i = 1,...,2000, we fin the list x t(i,1),...,x t(i,k) of k true nearest neighbors of x i, by irect scanning. Then, we compute the quantities Di true, D susp i, prop i, efine via (27), (28), (29), respectively, for i = 1,...,2000. We efine the average Dalgo true by the formula Dalgo true = Di true. (45) 2000 The averages D susp algo, prop algo are efine an compute by the same token. i=1 14

15 Generation of the points an invocation of RANN are repeate 20 times, to obtain the values Dalgo true (1),..., Dtrue algo (20), all efine via (45). Then, we efine the sample mean of via the formula D true algo E smpl [ D true ] = i=1 Dalgo true (i). (46) The sample means E smpl [D algo ] an E smpl [ propalgo ] are efine in a similar way. Finally, we compute the ratio ratio smpl = E smpl [D algo ] E smpl [D true ]. (47) The parameters for Experiment 2 were chosen in the following way. The number of points was N = The number of requeste nearest neighbors was k = 15, 30 or 60. The imensionality was chosen to be a multiple of 5 between 15 an 200. The number of iterations of the algorithm was either T = 1 or T = 10. The supercharging step was either skippe or performe once after T iterations. Most of the approximations have been compute with relative error up to 2%. The results are shown in Figures 2, 3 an in Table 1. In Figures 2, 3 (left), we plot ratio smpl (see (47)) as a function of the imensionality for k = 15 an k = 60, respectively. In Figures 2, 3 (right), we plot E smpl [ propalgo ] as a function of the imensionality for k = 15 an k = 60, respectively. Each figure contains four curves, corresponing to one iteration without supercharging, one iteration with supercharging, ten iterations without supercharging an ten iterations with supercharging. Table 1 has the following structure. The first column contains the imensionality. The next three columns contain the CPU time of 10 iterations of RANN without supercharging, with the requeste number of nearest neighbors being k = 15, 30, 60, respectively. The last three columns contain the CPU time of the supercharging only (after 10 iterations of RANN), with the requeste number of nearest neighbors being k = 15, 30, 60, respectively. The CPU time is shown in secons. Several observations can be mae from Table 1 an Figures 2, 3 an from more etaile experiments by the authors. In these observations, we refer to ratio smpl as ratio, an to E smpl [ propalgo ] as proportion. We recall that, roughly speaking, the proportion measures how many of the true nearest neighbors have been foun by RANN. On the other han, the ratio measures how much average istances to suspects iffer from the average istances to true nearest neighbors. 1. As expecte, for a fixe the performance of RANN improves as the number T of iterations increases, both in terms of ratio an proportion. 2. As expecte, for a fixe the performance of RANN improves if supercharging is performe, both in terms of ratio an proportion. 3. For a fixe, the performance of RANN improves as the number of requeste nearest neighbors k increases, both in terms of ratio an proportion (at the expense of running time). 15

16 4. For a fixe, the effects of supercharging (especially on proportion) increase as k grows. For example, in Figure 2, right (k = 15) we observe, that, for T = 10 an = 60, supercharging increases the proportion from 22% to 32%. On the other han, in Figure 3, right (k = 60) we observe that, for T = 10 an = 60, supercharging increases the proportion from 43% to 74%. 5. As expecte, the performance of RANN slowly eteriorates in terms of proportion, as increases. On the other han, there is no significant eterioration of performance in terms of ratio, as increases. 6. For T = 10 the ratio is always below 1.1, with or without supercharging. 7. Even for as high a imension as = 60, as few as ten iterations with supercharging correctly etermine at least 30% of the true nearest neighbors. Moreover, the error of this etection ecays exponentially with number of iterations T. 8. The running time of RANN, with or without supercharging, grows roughly linearly with imensionality, as expecte from (23) (see Table 1). 9. For fixe imensionality, the running time of RANN without supercharging grows roughly linearly with the requeste number of nearest neighbors k, as expecte from (23) (see Table 1). For example, 10 iterations of RANN in the case of = 200 take about 80, 124 an 208 secons for k = 15, 30, 60, respectively. 10. For fixe imensionality, the running time of the supercharging step grows roughly quaratically with the requeste number of nearest neighbors k, as expecte from (23) (see Table 1). For example, in the case of = 200 supercharging takes about 15, 57 an 235 secons for k = 15, 30, 60, respectively. 11. The algorithm has been teste on sets of points having non-gaussian istribution, e.g. the uniform istribution in the imensional cube [0, 1] or Hamming istribution (i.e. uniform istribution on the iscrete set of the vertices of [0, 1] ). For both uniform an Hamming istributions, the performance of the algorithm was very similar to that in the Gaussian case (see [10] for etails). 12. The algorithm has been teste on sets of points, having Gaussian istribution whose covariance matrix is not the ientity matrix. These tests seem to inicate that the performance of RANN improves as the conition number of the covariance matrix grows. As an extreme example, if the points belong to a q imensional linear subspace of R, the performance of the algorithm oes not epen on (though the running time obviously oes). References [1] A. Anoni, P. Inyk (2008) Near-Optimal Hashing Algorithms for Approximate Nearest Neighbor in High Dimensions, Communications of the ACM, 51(1):

17 10 iterations, without supercharging supercharging after 10 iterations k = 15 k = 30 k = 60 k = 15 k = 30 k = E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E+03 Table 1: CPU time of RANN, in secons. Number of points: 122,880. [2] S. Arya, D.M. Mount, N.S. Netanyahu, R. Silverman, A.Y. Wu (1998) An Optimal Algorithm for Approximate Nearest Neighbor Searching in Fixe Dimensions, Journal of the ACM, 45(6): [3] N. Ailon, B. Chazelle (2009) The Fast Johnson Linenstrauss Transform an Approximate Nearest Neighbors, SIAM J. Comput., 39(1): [4] N. Ailon, B. Chazelle (2010) Faster Dimension Reuction, Commun. ACM, 53(2): [5] W. Johnson, J. Linenstrauss (1984) Extensions of Lipschitz mappings into a Hilbert space, Contemporary Mathematics, 26: [6] V. Rokhlin, M. Tygert (2008) A fast ranomize algorithm for overetermine linear least-squares regression, Proc Natl Aca Sci USA 105(36): [7] E. Liberty, F. Woolfe, P.G. Martinsson, V. Rokhlin, M. Tygert (2007) Ranomize algorithms for the low-rank approximation of matrices, Proc Natl Aca Sci USA 104: [8] D. Knuth (1969) in Seminumerical Algorithms, vol. 2 of The Art of Computer Programming, Reaing, Mass: Aison-Wesley. [9] N. Ailon, E. Liberty (2010) Almost Optimal Unrestricte Fast Johnson-Linenstrauss Transform, eprint arxiv: [10] P.W. Jones, A. Osipov, V. Rokhlin (2010) A Ranomize Approximate Nearest Neighbors Algorithm, Yale technical report. 17

18 1.6 k = k = 30 D susp / D true proportion Figure 1: Number of points: N = (blue ots), (re squares), (black circles), (green pluses). k = 15, N = 122, k = 15, N = 122, D susp / D true proportion Figure 2: Without supercharging: one iteration (blue ots), ten iterations (black cirlces). With supercharging: one iteration (re squares), ten iterations (green pluses). k = 60, N = 122, k = 60, N = 122, D susp / D true proportion Figure 3: Without supercharging: one iteration (blue ots), ten iterations (black cirlces). With supercharging: one iteration (re squares), ten iterations (green pluses). 18

Lecture Introduction. 2 Examples of Measure Concentration. 3 The Johnson-Lindenstrauss Lemma. CS-621 Theory Gems November 28, 2012

Lecture Introduction. 2 Examples of Measure Concentration. 3 The Johnson-Lindenstrauss Lemma. CS-621 Theory Gems November 28, 2012 CS-6 Theory Gems November 8, 0 Lecture Lecturer: Alesaner Mąry Scribes: Alhussein Fawzi, Dorina Thanou Introuction Toay, we will briefly iscuss an important technique in probability theory measure concentration

More information

Lower bounds on Locality Sensitive Hashing

Lower bounds on Locality Sensitive Hashing Lower bouns on Locality Sensitive Hashing Rajeev Motwani Assaf Naor Rina Panigrahy Abstract Given a metric space (X, X ), c 1, r > 0, an p, q [0, 1], a istribution over mappings H : X N is calle a (r,

More information

Least-Squares Regression on Sparse Spaces

Least-Squares Regression on Sparse Spaces Least-Squares Regression on Sparse Spaces Yuri Grinberg, Mahi Milani Far, Joelle Pineau School of Computer Science McGill University Montreal, Canaa {ygrinb,mmilan1,jpineau}@cs.mcgill.ca 1 Introuction

More information

u!i = a T u = 0. Then S satisfies

u!i = a T u = 0. Then S satisfies Deterministic Conitions for Subspace Ientifiability from Incomplete Sampling Daniel L Pimentel-Alarcón, Nigel Boston, Robert D Nowak University of Wisconsin-Maison Abstract Consier an r-imensional subspace

More information

Lecture 6 : Dimensionality Reduction

Lecture 6 : Dimensionality Reduction CPS290: Algorithmic Founations of Data Science February 3, 207 Lecture 6 : Dimensionality Reuction Lecturer: Kamesh Munagala Scribe: Kamesh Munagala In this lecture, we will consier the roblem of maing

More information

Lower Bounds for the Smoothed Number of Pareto optimal Solutions

Lower Bounds for the Smoothed Number of Pareto optimal Solutions Lower Bouns for the Smoothe Number of Pareto optimal Solutions Tobias Brunsch an Heiko Röglin Department of Computer Science, University of Bonn, Germany brunsch@cs.uni-bonn.e, heiko@roeglin.org Abstract.

More information

Survey Sampling. 1 Design-based Inference. Kosuke Imai Department of Politics, Princeton University. February 19, 2013

Survey Sampling. 1 Design-based Inference. Kosuke Imai Department of Politics, Princeton University. February 19, 2013 Survey Sampling Kosuke Imai Department of Politics, Princeton University February 19, 2013 Survey sampling is one of the most commonly use ata collection methos for social scientists. We begin by escribing

More information

Implicit Differentiation

Implicit Differentiation Implicit Differentiation Thus far, the functions we have been concerne with have been efine explicitly. A function is efine explicitly if the output is given irectly in terms of the input. For instance,

More information

LECTURE NOTES ON DVORETZKY S THEOREM

LECTURE NOTES ON DVORETZKY S THEOREM LECTURE NOTES ON DVORETZKY S THEOREM STEVEN HEILMAN Abstract. We present the first half of the paper [S]. In particular, the results below, unless otherwise state, shoul be attribute to G. Schechtman.

More information

Diagonalization of Matrices Dr. E. Jacobs

Diagonalization of Matrices Dr. E. Jacobs Diagonalization of Matrices Dr. E. Jacobs One of the very interesting lessons in this course is how certain algebraic techniques can be use to solve ifferential equations. The purpose of these notes is

More information

Separation of Variables

Separation of Variables Physics 342 Lecture 1 Separation of Variables Lecture 1 Physics 342 Quantum Mechanics I Monay, January 25th, 2010 There are three basic mathematical tools we nee, an then we can begin working on the physical

More information

Computing Exact Confidence Coefficients of Simultaneous Confidence Intervals for Multinomial Proportions and their Functions

Computing Exact Confidence Coefficients of Simultaneous Confidence Intervals for Multinomial Proportions and their Functions Working Paper 2013:5 Department of Statistics Computing Exact Confience Coefficients of Simultaneous Confience Intervals for Multinomial Proportions an their Functions Shaobo Jin Working Paper 2013:5

More information

Math 342 Partial Differential Equations «Viktor Grigoryan

Math 342 Partial Differential Equations «Viktor Grigoryan Math 342 Partial Differential Equations «Viktor Grigoryan 6 Wave equation: solution In this lecture we will solve the wave equation on the entire real line x R. This correspons to a string of infinite

More information

The total derivative. Chapter Lagrangian and Eulerian approaches

The total derivative. Chapter Lagrangian and Eulerian approaches Chapter 5 The total erivative 51 Lagrangian an Eulerian approaches The representation of a flui through scalar or vector fiels means that each physical quantity uner consieration is escribe as a function

More information

Discrete Mathematics

Discrete Mathematics Discrete Mathematics 309 (009) 86 869 Contents lists available at ScienceDirect Discrete Mathematics journal homepage: wwwelseviercom/locate/isc Profile vectors in the lattice of subspaces Dániel Gerbner

More information

Table of Common Derivatives By David Abraham

Table of Common Derivatives By David Abraham Prouct an Quotient Rules: Table of Common Derivatives By Davi Abraham [ f ( g( ] = [ f ( ] g( + f ( [ g( ] f ( = g( [ f ( ] g( g( f ( [ g( ] Trigonometric Functions: sin( = cos( cos( = sin( tan( = sec

More information

LATTICE-BASED D-OPTIMUM DESIGN FOR FOURIER REGRESSION

LATTICE-BASED D-OPTIMUM DESIGN FOR FOURIER REGRESSION The Annals of Statistics 1997, Vol. 25, No. 6, 2313 2327 LATTICE-BASED D-OPTIMUM DESIGN FOR FOURIER REGRESSION By Eva Riccomagno, 1 Rainer Schwabe 2 an Henry P. Wynn 1 University of Warwick, Technische

More information

A new proof of the sharpness of the phase transition for Bernoulli percolation on Z d

A new proof of the sharpness of the phase transition for Bernoulli percolation on Z d A new proof of the sharpness of the phase transition for Bernoulli percolation on Z Hugo Duminil-Copin an Vincent Tassion October 8, 205 Abstract We provie a new proof of the sharpness of the phase transition

More information

Robust Forward Algorithms via PAC-Bayes and Laplace Distributions. ω Q. Pr (y(ω x) < 0) = Pr A k

Robust Forward Algorithms via PAC-Bayes and Laplace Distributions. ω Q. Pr (y(ω x) < 0) = Pr A k A Proof of Lemma 2 B Proof of Lemma 3 Proof: Since the support of LL istributions is R, two such istributions are equivalent absolutely continuous with respect to each other an the ivergence is well-efine

More information

PDE Notes, Lecture #11

PDE Notes, Lecture #11 PDE Notes, Lecture # from Professor Jalal Shatah s Lectures Febuary 9th, 2009 Sobolev Spaces Recall that for u L loc we can efine the weak erivative Du by Du, φ := udφ φ C0 If v L loc such that Du, φ =

More information

Lecture XII. where Φ is called the potential function. Let us introduce spherical coordinates defined through the relations

Lecture XII. where Φ is called the potential function. Let us introduce spherical coordinates defined through the relations Lecture XII Abstract We introuce the Laplace equation in spherical coorinates an apply the metho of separation of variables to solve it. This will generate three linear orinary secon orer ifferential equations:

More information

Time-of-Arrival Estimation in Non-Line-Of-Sight Environments

Time-of-Arrival Estimation in Non-Line-Of-Sight Environments 2 Conference on Information Sciences an Systems, The Johns Hopkins University, March 2, 2 Time-of-Arrival Estimation in Non-Line-Of-Sight Environments Sinan Gezici, Hisashi Kobayashi an H. Vincent Poor

More information

Linear First-Order Equations

Linear First-Order Equations 5 Linear First-Orer Equations Linear first-orer ifferential equations make up another important class of ifferential equations that commonly arise in applications an are relatively easy to solve (in theory)

More information

Thermal conductivity of graded composites: Numerical simulations and an effective medium approximation

Thermal conductivity of graded composites: Numerical simulations and an effective medium approximation JOURNAL OF MATERIALS SCIENCE 34 (999)5497 5503 Thermal conuctivity of grae composites: Numerical simulations an an effective meium approximation P. M. HUI Department of Physics, The Chinese University

More information

A Second Time Dimension, Hidden in Plain Sight

A Second Time Dimension, Hidden in Plain Sight A Secon Time Dimension, Hien in Plain Sight Brett A Collins. In this paper I postulate the existence of a secon time imension, making five imensions, three space imensions an two time imensions. I will

More information

6 General properties of an autonomous system of two first order ODE

6 General properties of an autonomous system of two first order ODE 6 General properties of an autonomous system of two first orer ODE Here we embark on stuying the autonomous system of two first orer ifferential equations of the form ẋ 1 = f 1 (, x 2 ), ẋ 2 = f 2 (, x

More information

Agmon Kolmogorov Inequalities on l 2 (Z d )

Agmon Kolmogorov Inequalities on l 2 (Z d ) Journal of Mathematics Research; Vol. 6, No. ; 04 ISSN 96-9795 E-ISSN 96-9809 Publishe by Canaian Center of Science an Eucation Agmon Kolmogorov Inequalities on l (Z ) Arman Sahovic Mathematics Department,

More information

On combinatorial approaches to compressed sensing

On combinatorial approaches to compressed sensing On combinatorial approaches to compresse sensing Abolreza Abolhosseini Moghaam an Hayer Raha Department of Electrical an Computer Engineering, Michigan State University, East Lansing, MI, U.S. Emails:{abolhos,raha}@msu.eu

More information

Influence of weight initialization on multilayer perceptron performance

Influence of weight initialization on multilayer perceptron performance Influence of weight initialization on multilayer perceptron performance M. Karouia (1,2) T. Denœux (1) R. Lengellé (1) (1) Université e Compiègne U.R.A. CNRS 817 Heuiasyc BP 649 - F-66 Compiègne ceex -

More information

The derivative of a function f(x) is another function, defined in terms of a limiting expression: f(x + δx) f(x)

The derivative of a function f(x) is another function, defined in terms of a limiting expression: f(x + δx) f(x) Y. D. Chong (2016) MH2801: Complex Methos for the Sciences 1. Derivatives The erivative of a function f(x) is another function, efine in terms of a limiting expression: f (x) f (x) lim x δx 0 f(x + δx)

More information

Examining Geometric Integration for Propagating Orbit Trajectories with Non-Conservative Forcing

Examining Geometric Integration for Propagating Orbit Trajectories with Non-Conservative Forcing Examining Geometric Integration for Propagating Orbit Trajectories with Non-Conservative Forcing Course Project for CDS 05 - Geometric Mechanics John M. Carson III California Institute of Technology June

More information

A fast randomized algorithm for overdetermined linear least-squares regression

A fast randomized algorithm for overdetermined linear least-squares regression A fast randomized algorithm for overdetermined linear least-squares regression Vladimir Rokhlin and Mark Tygert Technical Report YALEU/DCS/TR-1403 April 28, 2008 Abstract We introduce a randomized algorithm

More information

Schrödinger s equation.

Schrödinger s equation. Physics 342 Lecture 5 Schröinger s Equation Lecture 5 Physics 342 Quantum Mechanics I Wenesay, February 3r, 2010 Toay we iscuss Schröinger s equation an show that it supports the basic interpretation of

More information

ensembles When working with density operators, we can use this connection to define a generalized Bloch vector: v x Tr x, v y Tr y

ensembles When working with density operators, we can use this connection to define a generalized Bloch vector: v x Tr x, v y Tr y Ph195a lecture notes, 1/3/01 Density operators for spin- 1 ensembles So far in our iscussion of spin- 1 systems, we have restricte our attention to the case of pure states an Hamiltonian evolution. Toay

More information

Sturm-Liouville Theory

Sturm-Liouville Theory LECTURE 5 Sturm-Liouville Theory In the three preceing lectures I emonstrate the utility of Fourier series in solving PDE/BVPs. As we ll now see, Fourier series are just the tip of the iceberg of the theory

More information

Robustness and Perturbations of Minimal Bases

Robustness and Perturbations of Minimal Bases Robustness an Perturbations of Minimal Bases Paul Van Dooren an Froilán M Dopico December 9, 2016 Abstract Polynomial minimal bases of rational vector subspaces are a classical concept that plays an important

More information

Sharp Thresholds. Zachary Hamaker. March 15, 2010

Sharp Thresholds. Zachary Hamaker. March 15, 2010 Sharp Threshols Zachary Hamaker March 15, 2010 Abstract The Kolmogorov Zero-One law states that for tail events on infinite-imensional probability spaces, the probability must be either zero or one. Behavior

More information

Qubit channels that achieve capacity with two states

Qubit channels that achieve capacity with two states Qubit channels that achieve capacity with two states Dominic W. Berry Department of Physics, The University of Queenslan, Brisbane, Queenslan 4072, Australia Receive 22 December 2004; publishe 22 March

More information

Acute sets in Euclidean spaces

Acute sets in Euclidean spaces Acute sets in Eucliean spaces Viktor Harangi April, 011 Abstract A finite set H in R is calle an acute set if any angle etermine by three points of H is acute. We examine the maximal carinality α() of

More information

Function Spaces. 1 Hilbert Spaces

Function Spaces. 1 Hilbert Spaces Function Spaces A function space is a set of functions F that has some structure. Often a nonparametric regression function or classifier is chosen to lie in some function space, where the assume structure

More information

Make graph of g by adding c to the y-values. on the graph of f by c. multiplying the y-values. even-degree polynomial. graph goes up on both sides

Make graph of g by adding c to the y-values. on the graph of f by c. multiplying the y-values. even-degree polynomial. graph goes up on both sides Reference 1: Transformations of Graphs an En Behavior of Polynomial Graphs Transformations of graphs aitive constant constant on the outsie g(x) = + c Make graph of g by aing c to the y-values on the graph

More information

Database-friendly Random Projections

Database-friendly Random Projections Database-frienly Ranom Projections Dimitris Achlioptas Microsoft ABSTRACT A classic result of Johnson an Linenstrauss asserts that any set of n points in -imensional Eucliean space can be embee into k-imensional

More information

Math Notes on differentials, the Chain Rule, gradients, directional derivative, and normal vectors

Math Notes on differentials, the Chain Rule, gradients, directional derivative, and normal vectors Math 18.02 Notes on ifferentials, the Chain Rule, graients, irectional erivative, an normal vectors Tangent plane an linear approximation We efine the partial erivatives of f( xy, ) as follows: f f( x+

More information

Quantum Mechanics in Three Dimensions

Quantum Mechanics in Three Dimensions Physics 342 Lecture 20 Quantum Mechanics in Three Dimensions Lecture 20 Physics 342 Quantum Mechanics I Monay, March 24th, 2008 We begin our spherical solutions with the simplest possible case zero potential.

More information

Iterated Point-Line Configurations Grow Doubly-Exponentially

Iterated Point-Line Configurations Grow Doubly-Exponentially Iterate Point-Line Configurations Grow Doubly-Exponentially Joshua Cooper an Mark Walters July 9, 008 Abstract Begin with a set of four points in the real plane in general position. A to this collection

More information

On the Surprising Behavior of Distance Metrics in High Dimensional Space

On the Surprising Behavior of Distance Metrics in High Dimensional Space On the Surprising Behavior of Distance Metrics in High Dimensional Space Charu C. Aggarwal, Alexaner Hinneburg 2, an Daniel A. Keim 2 IBM T. J. Watson Research Center Yortown Heights, NY 0598, USA. charu@watson.ibm.com

More information

Tractability results for weighted Banach spaces of smooth functions

Tractability results for weighted Banach spaces of smooth functions Tractability results for weighte Banach spaces of smooth functions Markus Weimar Mathematisches Institut, Universität Jena Ernst-Abbe-Platz 2, 07740 Jena, Germany email: markus.weimar@uni-jena.e March

More information

Convergence of Random Walks

Convergence of Random Walks Chapter 16 Convergence of Ranom Walks This lecture examines the convergence of ranom walks to the Wiener process. This is very important both physically an statistically, an illustrates the utility of

More information

TOEPLITZ AND POSITIVE SEMIDEFINITE COMPLETION PROBLEM FOR CYCLE GRAPH

TOEPLITZ AND POSITIVE SEMIDEFINITE COMPLETION PROBLEM FOR CYCLE GRAPH English NUMERICAL MATHEMATICS Vol14, No1 Series A Journal of Chinese Universities Feb 2005 TOEPLITZ AND POSITIVE SEMIDEFINITE COMPLETION PROBLEM FOR CYCLE GRAPH He Ming( Λ) Michael K Ng(Ξ ) Abstract We

More information

SYNCHRONOUS SEQUENTIAL CIRCUITS

SYNCHRONOUS SEQUENTIAL CIRCUITS CHAPTER SYNCHRONOUS SEUENTIAL CIRCUITS Registers an counters, two very common synchronous sequential circuits, are introuce in this chapter. Register is a igital circuit for storing information. Contents

More information

Introduction to the Vlasov-Poisson system

Introduction to the Vlasov-Poisson system Introuction to the Vlasov-Poisson system Simone Calogero 1 The Vlasov equation Consier a particle with mass m > 0. Let x(t) R 3 enote the position of the particle at time t R an v(t) = ẋ(t) = x(t)/t its

More information

arxiv: v1 [cs.ds] 31 May 2017

arxiv: v1 [cs.ds] 31 May 2017 Succinct Partial Sums an Fenwick Trees Philip Bille, Aners Roy Christiansen, Nicola Prezza, an Freerik Rye Skjoljensen arxiv:1705.10987v1 [cs.ds] 31 May 2017 Technical University of Denmark, DTU Compute,

More information

d dx But have you ever seen a derivation of these results? We ll prove the first result below. cos h 1

d dx But have you ever seen a derivation of these results? We ll prove the first result below. cos h 1 Lecture 5 Some ifferentiation rules Trigonometric functions (Relevant section from Stewart, Seventh Eition: Section 3.3) You all know that sin = cos cos = sin. () But have you ever seen a erivation of

More information

Analysis on a Localized Pruning Method for Connected Dominating Sets

Analysis on a Localized Pruning Method for Connected Dominating Sets JOURNAL OF INFORMATION SCIENCE AND ENGINEERING 23, 1073-1086 (2007) Analysis on a Localize Pruning Metho for Connecte Dominating Sets JOHN SUM 1, JIE WU 2 AND KEVIN HO 3 1 Department of Information Management

More information

Leaving Randomness to Nature: d-dimensional Product Codes through the lens of Generalized-LDPC codes

Leaving Randomness to Nature: d-dimensional Product Codes through the lens of Generalized-LDPC codes Leaving Ranomness to Nature: -Dimensional Prouct Coes through the lens of Generalize-LDPC coes Tavor Baharav, Kannan Ramchanran Dept. of Electrical Engineering an Computer Sciences, U.C. Berkeley {tavorb,

More information

Pure Further Mathematics 1. Revision Notes

Pure Further Mathematics 1. Revision Notes Pure Further Mathematics Revision Notes June 20 2 FP JUNE 20 SDB Further Pure Complex Numbers... 3 Definitions an arithmetical operations... 3 Complex conjugate... 3 Properties... 3 Complex number plane,

More information

arxiv: v4 [cs.ds] 7 Mar 2014

arxiv: v4 [cs.ds] 7 Mar 2014 Analysis of Agglomerative Clustering Marcel R. Ackermann Johannes Blömer Daniel Kuntze Christian Sohler arxiv:101.697v [cs.ds] 7 Mar 01 Abstract The iameter k-clustering problem is the problem of partitioning

More information

A Weak First Digit Law for a Class of Sequences

A Weak First Digit Law for a Class of Sequences International Mathematical Forum, Vol. 11, 2016, no. 15, 67-702 HIKARI Lt, www.m-hikari.com http://x.oi.org/10.1288/imf.2016.6562 A Weak First Digit Law for a Class of Sequences M. A. Nyblom School of

More information

Estimation of the Maximum Domination Value in Multi-Dimensional Data Sets

Estimation of the Maximum Domination Value in Multi-Dimensional Data Sets Proceeings of the 4th East-European Conference on Avances in Databases an Information Systems ADBIS) 200 Estimation of the Maximum Domination Value in Multi-Dimensional Data Sets Eleftherios Tiakas, Apostolos.

More information

The Press-Schechter mass function

The Press-Schechter mass function The Press-Schechter mass function To state the obvious: It is important to relate our theories to what we can observe. We have looke at linear perturbation theory, an we have consiere a simple moel for

More information

Part I: Web Structure Mining Chapter 1: Information Retrieval and Web Search

Part I: Web Structure Mining Chapter 1: Information Retrieval and Web Search Part I: Web Structure Mining Chapter : Information Retrieval an Web Search The Web Challenges Crawling the Web Inexing an Keywor Search Evaluating Search Quality Similarity Search The Web Challenges Tim

More information

Analyzing Tensor Power Method Dynamics in Overcomplete Regime

Analyzing Tensor Power Method Dynamics in Overcomplete Regime Journal of Machine Learning Research 18 (2017) 1-40 Submitte 9/15; Revise 11/16; Publishe 4/17 Analyzing Tensor Power Metho Dynamics in Overcomplete Regime Animashree Ananumar Department of Electrical

More information

Permanent vs. Determinant

Permanent vs. Determinant Permanent vs. Determinant Frank Ban Introuction A major problem in theoretical computer science is the Permanent vs. Determinant problem. It asks: given an n by n matrix of ineterminates A = (a i,j ) an

More information

Multi-View Clustering via Canonical Correlation Analysis

Multi-View Clustering via Canonical Correlation Analysis Technical Report TTI-TR-2008-5 Multi-View Clustering via Canonical Correlation Analysis Kamalika Chauhuri UC San Diego Sham M. Kakae Toyota Technological Institute at Chicago ABSTRACT Clustering ata in

More information

Generalized Tractability for Multivariate Problems

Generalized Tractability for Multivariate Problems Generalize Tractability for Multivariate Problems Part II: Linear Tensor Prouct Problems, Linear Information, an Unrestricte Tractability Michael Gnewuch Department of Computer Science, University of Kiel,

More information

Diophantine Approximations: Examining the Farey Process and its Method on Producing Best Approximations

Diophantine Approximations: Examining the Farey Process and its Method on Producing Best Approximations Diophantine Approximations: Examining the Farey Process an its Metho on Proucing Best Approximations Kelly Bowen Introuction When a person hears the phrase irrational number, one oes not think of anything

More information

Lagrangian and Hamiltonian Mechanics

Lagrangian and Hamiltonian Mechanics Lagrangian an Hamiltonian Mechanics.G. Simpson, Ph.. epartment of Physical Sciences an Engineering Prince George s Community College ecember 5, 007 Introuction In this course we have been stuying classical

More information

d-dimensional Arrangement Revisited

d-dimensional Arrangement Revisited -Dimensional Arrangement Revisite Daniel Rotter Jens Vygen Research Institute for Discrete Mathematics University of Bonn Revise version: April 5, 013 Abstract We revisit the -imensional arrangement problem

More information

Expected Value of Partial Perfect Information

Expected Value of Partial Perfect Information Expecte Value of Partial Perfect Information Mike Giles 1, Takashi Goa 2, Howar Thom 3 Wei Fang 1, Zhenru Wang 1 1 Mathematical Institute, University of Oxfor 2 School of Engineering, University of Tokyo

More information

Optimization of Geometries by Energy Minimization

Optimization of Geometries by Energy Minimization Optimization of Geometries by Energy Minimization by Tracy P. Hamilton Department of Chemistry University of Alabama at Birmingham Birmingham, AL 3594-140 hamilton@uab.eu Copyright Tracy P. Hamilton, 1997.

More information

Chapter 4. Electrostatics of Macroscopic Media

Chapter 4. Electrostatics of Macroscopic Media Chapter 4. Electrostatics of Macroscopic Meia 4.1 Multipole Expansion Approximate potentials at large istances 3 x' x' (x') x x' x x Fig 4.1 We consier the potential in the far-fiel region (see Fig. 4.1

More information

inflow outflow Part I. Regular tasks for MAE598/494 Task 1

inflow outflow Part I. Regular tasks for MAE598/494 Task 1 MAE 494/598, Fall 2016 Project #1 (Regular tasks = 20 points) Har copy of report is ue at the start of class on the ue ate. The rules on collaboration will be release separately. Please always follow the

More information

Perfect Matchings in Õ(n1.5 ) Time in Regular Bipartite Graphs

Perfect Matchings in Õ(n1.5 ) Time in Regular Bipartite Graphs Perfect Matchings in Õ(n1.5 ) Time in Regular Bipartite Graphs Ashish Goel Michael Kapralov Sanjeev Khanna Abstract We consier the well-stuie problem of fining a perfect matching in -regular bipartite

More information

Assignment 1. g i (x 1,..., x n ) dx i = 0. i=1

Assignment 1. g i (x 1,..., x n ) dx i = 0. i=1 Assignment 1 Golstein 1.4 The equations of motion for the rolling isk are special cases of general linear ifferential equations of constraint of the form g i (x 1,..., x n x i = 0. i=1 A constraint conition

More information

JUST THE MATHS UNIT NUMBER DIFFERENTIATION 2 (Rates of change) A.J.Hobson

JUST THE MATHS UNIT NUMBER DIFFERENTIATION 2 (Rates of change) A.J.Hobson JUST THE MATHS UNIT NUMBER 10.2 DIFFERENTIATION 2 (Rates of change) by A.J.Hobson 10.2.1 Introuction 10.2.2 Average rates of change 10.2.3 Instantaneous rates of change 10.2.4 Derivatives 10.2.5 Exercises

More information

A Modification of the Jarque-Bera Test. for Normality

A Modification of the Jarque-Bera Test. for Normality Int. J. Contemp. Math. Sciences, Vol. 8, 01, no. 17, 84-85 HIKARI Lt, www.m-hikari.com http://x.oi.org/10.1988/ijcms.01.9106 A Moification of the Jarque-Bera Test for Normality Moawa El-Fallah Ab El-Salam

More information

and from it produce the action integral whose variation we set to zero:

and from it produce the action integral whose variation we set to zero: Lagrange Multipliers Monay, 6 September 01 Sometimes it is convenient to use reunant coorinates, an to effect the variation of the action consistent with the constraints via the metho of Lagrange unetermine

More information

TEMPORAL AND TIME-FREQUENCY CORRELATION-BASED BLIND SOURCE SEPARATION METHODS. Yannick DEVILLE

TEMPORAL AND TIME-FREQUENCY CORRELATION-BASED BLIND SOURCE SEPARATION METHODS. Yannick DEVILLE TEMPORAL AND TIME-FREQUENCY CORRELATION-BASED BLIND SOURCE SEPARATION METHODS Yannick DEVILLE Université Paul Sabatier Laboratoire Acoustique, Métrologie, Instrumentation Bât. 3RB2, 8 Route e Narbonne,

More information

arxiv: v1 [hep-lat] 19 Nov 2013

arxiv: v1 [hep-lat] 19 Nov 2013 HU-EP-13/69 SFB/CPP-13-98 DESY 13-225 Applicability of Quasi-Monte Carlo for lattice systems arxiv:1311.4726v1 [hep-lat] 19 ov 2013, a,b Tobias Hartung, c Karl Jansen, b Hernan Leovey, Anreas Griewank

More information

'HVLJQ &RQVLGHUDWLRQ LQ 0DWHULDO 6HOHFWLRQ 'HVLJQ 6HQVLWLYLW\,1752'8&7,21

'HVLJQ &RQVLGHUDWLRQ LQ 0DWHULDO 6HOHFWLRQ 'HVLJQ 6HQVLWLYLW\,1752'8&7,21 Large amping in a structural material may be either esirable or unesirable, epening on the engineering application at han. For example, amping is a esirable property to the esigner concerne with limiting

More information

4.2 First Differentiation Rules; Leibniz Notation

4.2 First Differentiation Rules; Leibniz Notation .. FIRST DIFFERENTIATION RULES; LEIBNIZ NOTATION 307. First Differentiation Rules; Leibniz Notation In this section we erive rules which let us quickly compute the erivative function f (x) for any polynomial

More information

Multidimensional Fast Gauss Transforms by Chebyshev Expansions

Multidimensional Fast Gauss Transforms by Chebyshev Expansions Multiimensional Fast Gauss Transforms by Chebyshev Expansions Johannes Tausch an Alexaner Weckiewicz May 9, 009 Abstract A new version of the fast Gauss transform FGT) is introuce which is base on a truncate

More information

Beating CountSketch for Heavy Hitters in Insertion Streams

Beating CountSketch for Heavy Hitters in Insertion Streams Beating CountSketch for eavy itters in Insertion Streams ABSTRACT Vlaimir Braverman Johns opkins University Baltimore, MD, USA vova@cs.jhu.eu Nikita Ivkin Johns opkins University Baltimore, MD, USA nivkin1@jhu.eu

More information

Robust Low Rank Kernel Embeddings of Multivariate Distributions

Robust Low Rank Kernel Embeddings of Multivariate Distributions Robust Low Rank Kernel Embeings of Multivariate Distributions Le Song, Bo Dai College of Computing, Georgia Institute of Technology lsong@cc.gatech.eu, boai@gatech.eu Abstract Kernel embeing of istributions

More information

Level Construction of Decision Trees in a Partition-based Framework for Classification

Level Construction of Decision Trees in a Partition-based Framework for Classification Level Construction of Decision Trees in a Partition-base Framework for Classification Y.Y. Yao, Y. Zhao an J.T. Yao Department of Computer Science, University of Regina Regina, Saskatchewan, Canaa S4S

More information

Monte Carlo Methods with Reduced Error

Monte Carlo Methods with Reduced Error Monte Carlo Methos with Reuce Error As has been shown, the probable error in Monte Carlo algorithms when no information about the smoothness of the function is use is Dξ r N = c N. It is important for

More information

Necessary and Sufficient Conditions for Sketched Subspace Clustering

Necessary and Sufficient Conditions for Sketched Subspace Clustering Necessary an Sufficient Conitions for Sketche Subspace Clustering Daniel Pimentel-Alarcón, Laura Balzano 2, Robert Nowak University of Wisconsin-Maison, 2 University of Michigan-Ann Arbor Abstract This

More information

Approximate Constraint Satisfaction Requires Large LP Relaxations

Approximate Constraint Satisfaction Requires Large LP Relaxations Approximate Constraint Satisfaction Requires Large LP Relaxations oah Fleming April 19, 2018 Linear programming is a very powerful tool for attacking optimization problems. Techniques such as the ellipsoi

More information

arxiv: v1 [math.mg] 10 Apr 2018

arxiv: v1 [math.mg] 10 Apr 2018 ON THE VOLUME BOUND IN THE DVORETZKY ROGERS LEMMA FERENC FODOR, MÁRTON NASZÓDI, AND TAMÁS ZARNÓCZ arxiv:1804.03444v1 [math.mg] 10 Apr 2018 Abstract. The classical Dvoretzky Rogers lemma provies a eterministic

More information

A Unified Theorem on SDP Rank Reduction

A Unified Theorem on SDP Rank Reduction A Unifie heorem on SDP Ran Reuction Anthony Man Cho So, Yinyu Ye, Jiawei Zhang November 9, 006 Abstract We consier the problem of fining a low ran approximate solution to a system of linear equations in

More information

Multi-View Clustering via Canonical Correlation Analysis

Multi-View Clustering via Canonical Correlation Analysis Keywors: multi-view learning, clustering, canonical correlation analysis Abstract Clustering ata in high-imensions is believe to be a har problem in general. A number of efficient clustering algorithms

More information

Lower Bounds for Local Monotonicity Reconstruction from Transitive-Closure Spanners

Lower Bounds for Local Monotonicity Reconstruction from Transitive-Closure Spanners Lower Bouns for Local Monotonicity Reconstruction from Transitive-Closure Spanners Arnab Bhattacharyya Elena Grigorescu Mahav Jha Kyomin Jung Sofya Raskhonikova Davi P. Wooruff Abstract Given a irecte

More information

Similarity Measures for Categorical Data A Comparative Study. Technical Report

Similarity Measures for Categorical Data A Comparative Study. Technical Report Similarity Measures for Categorical Data A Comparative Stuy Technical Report Department of Computer Science an Engineering University of Minnesota 4-92 EECS Builing 200 Union Street SE Minneapolis, MN

More information

Optimal Control of Spatially Distributed Systems

Optimal Control of Spatially Distributed Systems Optimal Control of Spatially Distribute Systems Naer Motee an Ali Jababaie Abstract In this paper, we stuy the structural properties of optimal control of spatially istribute systems. Such systems consist

More information

Calculus of Variations

Calculus of Variations 16.323 Lecture 5 Calculus of Variations Calculus of Variations Most books cover this material well, but Kirk Chapter 4 oes a particularly nice job. x(t) x* x*+ αδx (1) x*- αδx (1) αδx (1) αδx (1) t f t

More information

Well-posedness of hyperbolic Initial Boundary Value Problems

Well-posedness of hyperbolic Initial Boundary Value Problems Well-poseness of hyperbolic Initial Bounary Value Problems Jean-François Coulombel CNRS & Université Lille 1 Laboratoire e mathématiques Paul Painlevé Cité scientifique 59655 VILLENEUVE D ASCQ CEDEX, France

More information

G j dq i + G j. q i. = a jt. and

G j dq i + G j. q i. = a jt. and Lagrange Multipliers Wenesay, 8 September 011 Sometimes it is convenient to use reunant coorinates, an to effect the variation of the action consistent with the constraints via the metho of Lagrange unetermine

More information

Construction of the Electronic Radial Wave Functions and Probability Distributions of Hydrogen-like Systems

Construction of the Electronic Radial Wave Functions and Probability Distributions of Hydrogen-like Systems Construction of the Electronic Raial Wave Functions an Probability Distributions of Hyrogen-like Systems Thomas S. Kuntzleman, Department of Chemistry Spring Arbor University, Spring Arbor MI 498 tkuntzle@arbor.eu

More information

Introduction to variational calculus: Lecture notes 1

Introduction to variational calculus: Lecture notes 1 October 10, 2006 Introuction to variational calculus: Lecture notes 1 Ewin Langmann Mathematical Physics, KTH Physics, AlbaNova, SE-106 91 Stockholm, Sween Abstract I give an informal summary of variational

More information

Quantum Algorithms: Problem Set 1

Quantum Algorithms: Problem Set 1 Quantum Algorithms: Problem Set 1 1. The Bell basis is + = 1 p ( 00i + 11i) = 1 p ( 00i 11i) + = 1 p ( 01i + 10i) = 1 p ( 01i 10i). This is an orthonormal basis for the state space of two qubits. It is

More information