Database-friendly Random Projections

Size: px
Start display at page:

Download "Database-friendly Random Projections"

Transcription

1 Database-frienly Ranom Projections Dimitris Achlioptas Microsoft ABSTRACT A classic result of Johnson an Linenstrauss asserts that any set of n points in -imensional Eucliean space can be embee into k-imensional Eucliean space where k is logarithmic in n an inepenent of so that all pairwise istances are maintaine within an arbitrarily small factor. All known constructions of such embeings involve projecting the n points onto a ranom k-imensional hyperplane. We give a novel construction of the embeing, suitable for atabase applications, which amounts to computing a simple aggregate over k ranom attribute partitions.. INTRODUCTION Consier projecting the points of your favorite sculpture first onto a plane an then onto a single line. The result amply emonstrates the power of imensionality. Conversely, given a high-imensional pointset it is natural to ask whether it exploits its full allotment of imensionality or, rather, it coul be embee into a lower imensional space without suffering great istortion. In general, such questions involve a, perhaps infinite, collection of points enowe with some istance function metric. In this paper, we will only eal with finite sets of points in Eucliean space so the Eucliean istance is the metric. In particular, it will be convenient to think of n points in R as an n table matrix A with each point represente as a row vector with attributes coorinates. Given such a matrix A, one of the most common embeings is the one suggeste by its Singular Value Decomposition. In particular, to embe the n points into R k we project them onto the k-imensional space spanne by the singular vectors corresponing to the k largest singular values of A. If one rewrites the result of this projection as a rank k Aress: Microsoft Corporation, One Microsoft Way, Remon WA, 98052, U.S.A. optas@microsoft.com n matrix A k, we are guarantee that for every rank k matrix D A A k A D, for any unitarily invariant norm, such as the Frobenius or the L2 norm. Thus, istortion here amounts to a certain istance norm between the set of projecte points, A k, an the original set of points A. If we associate with each row point a vector corresponing to the ifference between its original an its new position then, for example, uner the Frobenius norm the istortion equals the sum of the square lengths of these vectors. It is clear that such a notion of istortion captures a significant global property. At the same time, though, it oes not offer any local guarantees. For example, the istance between a pair of points can be arbitrarily smaller than what it was in the original space, if that is avantageous to minimizing the total istortion. The stuy of embeings that respect local properties is a rich area of mathematics with eep an beautiful results. Such embeings can guarantee, for example, that all istances between pairs of points are approximately maintaine or, more generally, that for a given q 2, a certain notion of volume is maintaine for all collections of up to q points thus capturing higher orer local structure. The algorithmic uses of such embeings were first consiere in the seminal paper of Linial, Lonon an Rabinovich 9] an have by now become an important part of moern algorithmic esign. A real gem in this area has been the following result of Johnson an Linenstrauss 7]. Lemma 7]. Given ɛ > 0 an an integer n, let k be a positive integer such that k k 0 Oɛ 2 log n. For every set P of n points in R there exists f : R R k such that for all u, v P ɛ u v 2 fu fv 2 + ɛ u v 2. We will refer to embeings proviing a guarantee akin to that of Lemma as JL-embeings. In the last few years, JL-embeings have been useful in solving a variety of problems. The rough iea is the following. By proviing a low imensional representation of the ata, JL-embeings spee up certain algorithms ramatically, in particular algorithms whose run-time epens exponentially in the imension of the working space there are a number of practical problems for which the best known algorithms have such behaviour. At the same time, the provie guarantee regaring pairwise istances is often enough to establish that the solution

2 foun by working in the low imensional space is a goo approximation to the optimal solution in the original space. We give a few examples below. Papaimitriou, Raghavan, Tamaki an Vempala 0], prove that embeing the points of A in a low-imensional space can significantly spee up the computation of a low rank approximation to A, without significantly affecting its quality. In 6], Inyk an Motwani showe that JL-embeings are useful in solving the ε-approximate nearest neighbor problem, where after some preprocessing of the pointset P one is to answer queries of the following type: Given an arbitrary point x, fin a point y P, such that for every point z P, x z ε x y. In a ifferent vein, Schulman ] use JL-embeings as part of an approximation algorithm for the version of clustering where we seek to minimize the sum of the squares of intracluster istances. Recently, Inyk 5] showe that JL-embeings can also be use in the context of ata-stream computation, where one has limite memory an is allowe only a single pass over the ata stream.. Our contribution Over the years, the probabilistic metho has allowe for the original proof of Johnson an Linenstrauss to be greatly simplifie an sharpene 4, 6, 3], while at the same time giving conceptually simple ranomize algorithms for constructing the embeing. Roughly speaking, all such algorithms project the input points onto a spherically ranom hyperplane through the origin. Performing such a projection, while conceptually simple, is non-trivial, especially in a atabase environment. Moreover, its computational cost can be prohibitive for certain applications. At the same time, JL-embeings have become an important algorithmic esign tool an in certain omains they are a esirable stanar ata processing step. With this in min, it is natural to ask if we can compute such embeings in a manner that is simpler an more efficient than the one suggeste by the current methos. Our main result, below, is a first step in this irection, asserting that one can replace projections onto ranom hyperplanes with much simpler an faster operations, requiring extremely simple probability istributions. In particular, these operations can be implemente reaily using stanar SQL primitives without any aitional functionality. Moreover, somewhat surprisingly, this comes without any sacrifice in the quality of the embeing. In fact, we will see that for every fixe value of we can get slightly better bouns than all current methos. We escribe the main result below in stanar mathematical terminology. Following that, we give an example of how to compute the embeing using atabase operations. As in Lemma, the parameter ɛ controls the accuracy in istance preservation, while now β controls the probability of success. Theorem 2. Let P be an arbitrary set of n points in R, represente as an n matrix A. Given ɛ, β > 0 let k β ɛ 2 /2 ɛ 3 /3 log n. For integer k k 0, let R be a k ranom matrix with Ri, j r ij, where {r ij} are inepenent ranom variables from either one of the following two probability istributions: { + with probability /2 r ij /2, r ij + with probability / /3 /6. Let E k A R. Let f : R R k map the i th row of A to the i th row of E. With probability at least n β, for all u, v P ɛ u v 2 fu fv 2 + ɛ u v 2. In a atabase system, all operations neee to compute A R are very efficient an easy to implement. For example, with the secon istribution above, the embeing amounts to generating k new attributes, each one forme by applying the same process: throw away 2/3 of all attributes at ranom; partition the remaining attributes ranomly into two equal parts; for each partition, prouce a new attribute equal to the sum of all attributes; take the ifference of the two sum-attributes. All in all, using Theorem 2, one nees very simple probability istributions, no floating point arithmetic, an all computation amounts to highly optimize atabase operations aggregation. By using the secon probability istribution, where r ij 0 with probability 2/3, we also get a threefol speeup as we only nee to process a thir of all attributes for each of the k coorinates. On the other han, when r ij {, +}, conceptually the construction seems to be about as simple as one coul hope for. Looking a bit more closely into the matrix E we see that each row vector of A is projecte onto k ranom vectors whose coorinates {r ij} are inepenent ranom variables with mean 0 an variance. If the {r ij} were inepenent Normal ranom variables with mean 0 an variance, it is well-known that the resulting vectors woul point to uniformly ranom irections in space. Projections onto such ranom lines through the origin have been consiere in a number of settings, incluing the work of Kleinberg on approximate nearest neighbors 8] an of Vempala on learning intersections of halfspaces 2]. More recently, such projections have also been use in learning mixture of Gaussians moels, starting with the work of Dasgupta 2] an later with the work of Arora an Kannan ]. Our proof will suggest that for any fixe vector α, the behavior of its projection onto a ranom vector c is manate by the even moments of α c. In fact, our result follows by showing that for every vector α, uner our istributions for {r ij}, these moments are ominate by the corresponing moments for the case where c is spherically symmetric. As a result, projecting onto vectors whose entries are istribute

3 like the columns of matrix R coul replace projection onto ranom lines; it is computationally simpler an results in projections that are at least as nicely behave. Finally, we note that Theorem 2 allows one to use significantly fewer ranom bits than all previous methos for constructing JL-embeings. While the amount of ranomness neee is still quite large, such attempts for ranomness reuction are of inepenent interest an our result can be viewe as a first step in that irection. 2. PREVIOUS WORK Let us write X D Y to enote that X is istribute as Y an recall that N0, enotes the stanar Normal ranom variable having mean 0 an variance. As we will see, in all methos for proucing JL-embeings, incluing ours, the heart of the matter is showing that for any vector, the square length of its projection is sharply concentrate aroun its expecte value. Arme with a sufficiently strong such concentration boun, one then proves the assertion of Lemma for a collection of n points in R by applying the union boun for the n 2 events corresponing to each istance-vector being istorte by more than ±ɛ. The original proof of Johnson an Linenstrauss 7] uses quite heavy geometric approximation machinery to yiel such a concentration boun when the projection is onto a uniformly ranom hyperplane through the origin. That proof was greatly simplifie an sharpene by Frankl an Meahara 4] who consiere a irect projection onto k ranom orthonormal vectors, yieling the following result. Theorem 3 4]. For any ɛ 0, /2, any sufficiently large set P R, an k k 0 9ɛ 2 2ɛ 3 /3 log P +, there exists a map f : P R k such that for all u, v P, ɛ u v 2 fu fv 2 + ɛ u v 2. The next great simplification of the proof of Lemma was given, inepenently, by Inyk an Motwani 6] an Dasgupta an Gupta 3], the latter also giving a slight sharpening of the boun for k 0. Below we state our renition of how this simplification was achieve. Assume that we try to implement the scheme of Frankl an Maehara 4] but we are lazy about enforcing either normality unit length or orthogonality among our k vectors. Instea, we just pick our k vectors inepenently, in a spherically symmetric manner. As we saw earlier, we can achieve this by taking as the coorinates of each vector inepenent N0, ranom variables. We then merely scale each vector by / so that its expecte length is. An immeiate gain of this approach is that now, for any fixe vector α, the length of its projection onto each of our vectors is also a Normal ranom variable. This is ue to a powerful an eep fact, namely the 2-stability of the Gaussian istribution: for any real numbers α, α 2,..., α, if {Z i} i is a family of inepenent Normal ranom variables an X i αizi, then X D c N0,, where c α 2 + +α 2 /2. As a result, if we interpret each of the k projection lengths as a coorinate in R k, then the square length of the resulting vector follows the Gamma istribution for which strong concentration bouns are reaily available. An what have we lost? Surprisingly little. While we i not insist upon either orthogonality, or normality, with high probability, the resulting k vectors come very close to having both these properties. In particular, the length of each of the k vectors is sharply concentrate aroun as the sum of inepenent ranom variables. Moreover, since the k vectors point in uniformly ranom irections in R, they get rapily closer to being orthogonal as grows. Unlike Inyk an Motwani 6], Dasgupta an Gupta 3] exploite spherical symmetry without appealing irectly to the 2-stability of the Gaussian istribution. Instea they observe that, by symmetry, the projection of any unit vector α on a ranom hyperplane through the origin is istribute exactly like the projection of a ranom point from the surface of the -imensional sphere onto a fixe subspace of imension k. Such a projection can be stuie reaily, though, as now each coorinate is a scale Normal ranom variable. Their analysis gave the strongest known boun, namely k k 0 4ɛ 2 /2 ɛ 3 /3. Note that this is exactly the same as our boun in Theorem 2 as β tens to SOME INTUITION By combining the analysis of 3] with the viewpoint of 6] it is in fact not har to show that Theorem 2 hols if for all i, j, r ij D N0,. Thus, our contribution essentially begins with the realization that spherical symmetry, while making life extremely comfortable, is not essential. What is essential is concentration. So, at least in principle, one is free to consier other caniate istributions for the {r ij}, if perhaps at the expense of comfort. As we saw earlier, each column of our matrix R will give us a coorinate of the projection in R k. Moreover, the square length of the projection is merely the sum of the squares of these coorinates. So, effectively, each column acts as an estimator of the original vector s length by taking its inner prouct with it, while in the en we take the consensus estimate sum over our k estimators. From this point of view, requiring our k vectors to be orthonormal has the pleasant statistical interpretation of greatest efficiency. In any case, though, as long as each column is an unbiase, boune variance estimator the Central Limit Theorem asserts that by taking enough columns we can get an arbitrarily goo estimate of the original length. Naturally, how many estimators are enough epens solely on the variance of the estimators. So, alreay we see that the key issue is the concentration of the projection of an arbitrary fixe vector α onto a single ranom vector. The main technical ifficulty that results from giving up spherical symmetry is that this concentration can now epen on α. Our main technical contribution lies in etermining probability istributions for {r ij} for which this concentration, for all vectors, is as goo as when r D ij N0,. In fact, it will turn out that for every fixe value of, we can get a minuscule improvement

4 over the concentration for that case. Thus, for every fixe, we can actually get a strictly better boun for k, albeit marginally, than by taking spherically ranom vectors. The reaer might be wonering how can it be that perfect spherical symmetry oes not buy us anything? an is in fact slightly worse for each fixe. At a high level, an answer to this question might go as follows. Given that we o not have spherical symmetry anymore, an aversary coul try to pick a vector α so that the length of its projection is as variable as possible. It is clear that not all vectors α are equal with respect to this variability. What then oes a worst-case vector w look like? How much are we exposing to the aversary by committing to pick our column vectors among lattice points rather than arbitrary points in R? As we will see, the worst-case vector is w /,..., an all 2 vectors resulting by sign-flipping w s coorinates. So, the worst-case vector turns out to be a more or less typical vector, at least in terms of the fluctuations in its coorinates, unlike say, 0,..., 0. As a result it is not har to believe that the aversary woul not fare much worse by picking a ranom vector. But in that case the aversary oes not benefit at all from our commitment. To get a more satisfactory answer, it seems like one has to elve into the proof. In particular, both for the spherically ranom case an for our istributions, the boun on k is manate by the probability of overestimating the projecte length. Thus, the ba events amount to the spanning vectors being too well-aligne with α. As a result, for any fixe one has to consier the traeoff between the probability an the extent of alignment. For example, let us consier the projection onto a single ranom vector when 2 an r ij {, +}. As we sai above, the worst case vector is w / 2,. So, it s easy to see that with probability /2 we have perfect alignment when our ranom vector is ±w an with probability /2 we have orthogonality. On the other han, for the spherically symmetric case, we have to consier the integral over all points on the plane, weighte by their probability uner the two-imensional Gaussian istribution. By a convexity argument it turns out that for every fixe, the even moments of the projecte length are marginally greater in the spherically symmetric case. This leas to a marginally weaker probability boun for that case. As one might guess, the two bouns coincie as tens to infinity. 4. PRELIMINARIES Let x y enote the inner prouct of vectors x, y. To simplify notation in the calculations, we will work with matrix R scale by /. Thus, R is a ranom k matrix with Ri, j r ij/, where the {r ij} are istribute as in Theorem 2. As a result, to get E we nee to scale A R by /k rather than / k. Therefore, if c j enotes the j th column of R, then {c j} k j is a family of k i.i.. ranom unit vectors in R an for all α R, fα /k α c,..., α c. In practice, of course, such scaling can be postpone until after the matrix multiplication projection has been performe, so that we maintain the avantage of only having {, 0, +} in the projection matrix. Let us start by computing E fα 2 for an arbitrary vector α R. Let {Q j} k j be efine as Q j α c j. Then E Q j E α ir ij i i an E 2 Q 2 j E α ir ij E i i α ir ij 2 + αi 2 E rij 2 + l m l m α ie r ij 0, i 2α l α mr lj r mj 2α l α me r lj E r mj α 2. 2 Note that to get an 2 we only use that {r ij} are inepenent, Er ij 0 an Varr ij. Using 2 we get E fα 2 k k E Q 2 j α 2. j That is, E fα 2 α 2 for any inepenent family of {r ij} with Er ij 0 an Varr ij. From the above we see that any istribution where Er ij 0 an Varr ij is, in principle, a caniate for the entries of R. In fact, in 3], Arriaga an Vempala inepenently suggeste the possibility of getting JL-embeings by projecting onto a matrix where r ij {, +} but i not give any bouns on the necessary value of k. As we mentione earlier, having a JL-embeing amounts to the following: for each of the n 2 pairs u, v P, the square norm of the vector u v, is maintaine within a factor of ± ɛ. Therefore, if we can prove that for some β > 0 an every vector α R, Pr ɛ α 2 fα 2 +ɛ α 2 ] 2, 3 n2+β then the probability that our projection oes not yiel a JL-embeing is boune by n 2 2/n 2+β < /n β. Let us note that since for a fixe projection matrix, fα 2 is proportional to α, it suffices to consier probability bouns for arbitrary unit vectors. Moreover, note that when E fα 2 α 2, inequality 3 merely asserts that the ranom variable fα 2 is concentrate aroun its expectation. Before consiering this point for our istributions for {r ij}, let us first wrap up the spherically ranom case. Getting a concentration inequality for fα 2 when r ij D N0, is straightforwar. Due to the 2-stability of the Normal istribution, fα 2 follows the Gamma istribution,

5 with parameters λ an t k, for every unit vector α. The fact that we get the same istribution for every vector α correspons to the intuition that all vectors are the same with respect to projection onto a spherically ranom vector. Stanar tail-bouns for the Gamma istribution reaily yiel the following. Lemma 4. Let r D ij N0, for all i, j. Then, for any ɛ > 0 an any unit-vector α R, Pr fα 2 + ɛk/ ] < exp k 2 ɛ2 /2 ɛ 3 /3, Pr fα 2 ɛk/ ] < exp k 2 ɛ2 /2 ɛ 3 /3. Thus, to get a JL-embeing we nee only require 2 exp k 2 ɛ2 /2 ɛ 3 /3 2 n, 2+β which hols for k 4 + 2β ɛ 2 /2 ɛ 3 /3 log n. We will use the stanar technique of applying Markov s inequality to the moment generating function of S. In particular, for arbitrary h > 0 we write Pr S > + ɛ k ] Pr exphs > exp h + ɛ k ] < E exp hs exp h + ɛ k Since {Q j} k j are i.i.. we get k E exp hs E exp hq 2 j 4 k j j E exp hq 2 j. 5 E exp hq 2 k, 6 where passing from 4 to 5 uses that the {Q j} k j are inepenent, while passing from 5 to 6 uses that they are ientically istribute. Thus, for any ɛ > 0 Pr S > + ɛ k ] < E exp hq 2 k exp h + ɛ k. 7 Let us note that the boun on the upper tail of fα 2 above is tight up to lower orer terms. As a result, as long as the union boun is use, one cannot hope for a better boun on k while using spherically ranom vectors. To prove our result we use the exact same approach, arguing that for every unit vector α R, the ranom variable αc 2 is sharply concentrate aroun its expectation, where c is a column of our projection matrix R. In the next section we state a lemma analogous to Lemma 4 above an show how it follows from bouns on certain moments of αc 2. We prove those bouns in Section PROBABILITY BOUNDS To simplify notation let us efine for an arbitrary vector α, S Sα k α c j 2 j k Q 2 jα, j where c j is the j th column of R, so that fα 2 S /k. Lemma 5. Let r ij have any of the two istributions in Theorem 2. Then, for any ɛ > 0 an any unit vector α R, Pr S > + ɛk/] < exp k 2 ɛ2 /2 ɛ 3 /3, Pr S < ɛk/] < exp k 2 ɛ2 /2 ɛ 3 /3. We will get a tight boun on E exp hq 2 from Lemma 6 below. Similarly, but this time consiering exp hs for arbitrary h > 0, we get that for any ɛ > 0 Pr S < ɛ k ] < E exp hq 2 k exp h ɛ k. 8 Rather than bouning E exp hq 2 irectly, this time we will expan exphq 2 to get Pr S < ɛ k ] 9 hq 2 2 k < E hq 2 + exp h ɛ k 2! h h2 + 2 E Q 4 k exp h ɛ k, 0 where EQ 2 was given by 2. We will get a tight boun on E Q 4 from Lemma 6 below. Lemma 6. For all h 0, /2 an all, E exp hq 2 2h/, In proving Lemma 5 we will generally omit the epenence of probabilities on α, making it explicit only when it affects our calculations. E Q The proof of Lemma 6 will comprise Section 6. Below we show how it implies Lemma 5 an thus Theorem 2.

6 Proof of Lemma 5. Substituting in 7 we get 3. To optimize the boun we set the erivative in 3 with respect to h to 0. This gives h ɛ <. Substituting 2 +ɛ 2 this value of h we get 4 an series expansion yiels 5. Pr S > + ɛ k ] k exp h + ɛ k 3 2h/ + ɛ exp ɛ k/2 4 < exp k 2 ɛ2 /2 ɛ 3 /3. 5 Similarly, substituting 2 in 8 we get 6. This time taking h ɛ is not optimal but it is goo enough, 2 +ɛ giving 7. Again, series expansion yiels 8. Pr S < ɛ k ] h k h exp h ɛ k 6 2 ɛ 2 + ɛ + 3ɛ 2 k ɛ ɛk exp ɛ ɛ < exp k 2 ɛ2 /2 ɛ 3 / MOMENT BOUNDS Here we prove bouns on certain moments of Q. To simplify notation, we rop the subscript, writing it as Q. It shoul be clear that the istribution of Q epens on α, i.e., Q Qα. This is precisely what we give up by not projecting onto spherically symmetric vectors. Our strategy for giving bouns on the moments of Q will be to etermine a worst case unit vector w an consier Qw. Our precise claim is the following. Lemma 7. Let w,...,. For every unit vector α R, an for all k 0,,... E Qα 2k E Qw 2k. 9 Moreover, we will prove that the even moments of Qw are ominate by the even moments of an appropriately scale Normal ranom variable, i.e., the corresponing moments from the spherically symmetric case. Lemma 8. Let T D N0, /. For all an all k 0,,... E Qw 2k E T 2k. 20 Postponing the proof of Lemmata 7 an 8 for a moment, let us use them to prove Lemma 6. Proof of Lemma 6. We start by observing that E T 4 + exp λ 2 λ 4 /2 λ 3 2π. 2 Along with 9 an 20 this reaily implies 2. For any real-value ranom variable U, the Monotone Convergence Theorem MCT implies E exp hu 2 hu 2 k h k E k! k! E U 2k k0 2 k0 for all h such that E exp hu 2 is boune. For E exp ht 2, below, taking h 0, /2 makes the integral converge, giving 2. Thus, for such h we can apply the MCT to get 22. Now, applying 9 an 20 to 22 gives 23. Applying the MCT once more gives 24. E exp ht 2 + 2π exp λ 2 /2 exp h λ2 λ 2h/ 2 k0 k0 h k k! E T 2k 22 h k k! E Q 2k 23 E exp hq Thus, E exp hq 2 / 2h/ for h 0, /2, which is precisely inequality. To prove Lemma 7 we nee the following lemma. It s proof appears in the Appenix. Lemma 9. Let r, r 2 be i.i.. r.v. having one of the following two probability istributions: r i {, +}, each value having probability /2, or, r i { 3, 0, + 3} with 0 having probability 2/3 an ± 3 being equiprobable. For real numbers a, b let c a 2 + b 2 /2. Then, for all T an all k 0,,... E T + ar + br 2 2k E T + cr + cr 2 2k. Proof of Lemma 7. Recall that for any vector α, Qα Q α α c where c r,..., r. If α α,..., α is such that αi 2 αj 2 for all i, j, then by symmetry, Qα an Qw are ientically istribute an the lemma hols trivially. Otherwise, we can assume without loss of generality, that α 2 α2 2 an consier the

7 more balance unit vector θ c, c, α 3,..., α, where c α 2 + α2 2 /2. We will prove that E Qα 2k E Qθ 2k. 25 Applying this argument repeately yiels the lemma, as θ eventually becomes w. To prove 25, below we first express E Qα 2k as a sum of averages over r, r 2. We then apply Lemma 9 to get that each term average in the sum, is boune by the corresponing average for vector θ. More precisely, E Qα 2k ] E R + α r k + α 2r 2 2k Pr α ir i R R R i3 i3 ] E R + cr k + cr 2 2k Pr α ir i R E Qθ 2k. To prove the lemma we will show that for every value assignment to the inices i,..., i 2k, E Y i Y i2k E T i T i2k. 26 Let V v, v 2,..., v 2k be the value assignment consiere. For i {,..., }, let C V i be the number of times that i appears in V. Observe that if for some i, c V i is o then both expectations appearing in 26 are 0, since both {Y i} i an {T i} i are inepenent families an EY i ET i 0 for all i. Thus, we can assume that there exists a set {j, j 2,..., j p} of inices an corresponing values l, l 2,..., l p such that E Y i Y i2k E E T i T i2k E Y 2l j Y 2l 2 j 2 Y 2lp j p, an T 2l j T 2l 2 j 2 T 2lp j p Note now that since the inices j, j 2,..., j p are istinct, {Y jt } p t an {Tj t} p t are families of i.i.. r.v. Therefore, E Y i Y i2k E Y 2l j E Y 2lp j p, an E T i T i2k E T 2l j E T 2lp j p.. Proof of Lemma 8. Recall that T D N0, /. We will first express T as the scale sum of inepenent stanar Normal ranom variables. This will allow for a irect comparison of the terms in each of the two expectations. Specifically, let {T i} i be a family of i.i.. stanar Normal ranom variables. Then i Ti is a Normal ranom variable with variance. Therefore, T D T i. i Recall also that Qw Q w w c where c r,..., r. To simplify notation let us write r i Y i an let us also rop the epenence of Q on w. Thus, Q Y i, i where {Y i} i are i.i.. r.v. having one of the following two istributions: Y i {, +}, each value having probability /2, or Y i { 3, 0, + 3} with 0 having probability 2/3 an ± 3 being equiprobable. We are now reay to compare E Q 2k with E T 2k. We first observe that for every k 0,,... E T 2k E Q 2k 2k 2k i i i 2k i 2k E T i T i2k, an E Y i Y i2k. So, without loss of generality, in orer to prove 26 it suffices to prove that for every l 0,,... E E. 27 Y 2l T 2l This, though, is completely trivial. Moreover, along with Lemma 9, it is the only point were we nee to use properties of the istribution for the r ij here calle Y i. Let us first recall the well-known fact that the 2lth moment of N0, is 2l!! 2l!/l!2 l. Furthermore: If Y {, +} then E Y 2l. If Y { 3, 0, + 3} then EY 2l 3 l 2l!/l!2 l, where the last inequality follows by an easy inuction. Let us note that since E Y 2l < E T 2l for certain l, one can get that for each fixe, both inequalities in Lemma 6 are actually strict, yieling slightly better tails bouns for S an a corresponingly better boun for k 0. As a last remark we note that by using Jensen s inequality one can get a irect boun for EQ 2k when Y i {, +}, i.e., without comparing it to ET 2k. That simplifies the proof for that case an shows that taking Y i {, +} is the minimizer of E exp hq 2 for all h. Acknowlegments I am grateful to Marek Biskup for his help with the proof of Lemma 8 an to Jeong Han Kim for suggesting the approach of equation 0. Many thanks also to Paul Braley, Anna Karlin, Elias Koutsoupias an Piotr Inyk for comments on earlier versions of the paper an useful iscussions. 7. REFERENCES ] S. Arora an R. Kannan. Learning mixtures of arbitrary Gaussians. Submitte, 2000.

8 2] S. Dasgupta. Learning mixtures of Gaussians. In 40th Annual Symposium on Founations of Computer Science New York, NY, 999, pages IEEE Comput. Soc. Press, Los Alamitos, CA, ] S. Dasgupta an A. Gupta. An elementary proof of the Johnson-Linenstrauss lemma. Technical report , UC Berkeley, March ] P. Frankl an H. Maehara. The Johnson-Linenstrauss lemma an the sphericity of some graphs. J. Combin. Theory Ser. B, 443: , ] P. Inyk. Stable istributions, pseuoranom generators, embeings an ata stream computation. In 4st Annual Symposium on Founations of Computer Science Reono Beach, CA, 2000, pages IEEE Comput. Soc. Press, Los Alamitos, CA, ] P. Inyk an R. Motwani. Approximate nearest neighbors: towars removing the curse of imensionality. In 30th Annual ACM Symposium on Theory of Computing Dallas, TX, pages ACM, New York, ] W. B. Johnson an J. Linenstrauss. Extensions of Lipschitz mappings into a Hilbert space. In Conference in moern analysis an probability New Haven, Conn., 982, pages Amer. Math. Soc., Provience, R.I., ] J. Kleinberg. Two algorithms for nearest-neighbor search in high imensions. In 29th Annual ACM Symposium on Theory of Computing El Paso, TX, 997, pages ACM, New York, ] N. Linial, E. Lonon, an Y. Rabinovich. The geometry of graphs an some of its algorithmic applications. Combinatorica, 52:25 245, ] C. H. Papaimitriou, P. Raghavan, H. Tamaki, an S. Vempala. Latent semantic inexing: A probabilistic analysis. In 7th Annual Symposium on Principles of Database Systems Seattle, WA, 998, pages 59 68, 998. ] L. J. Schulman. Clustering for ege-cost minimization. In 32n Annual ACM Symposium on Theory of Computing Portlan, OR, 2000, pages ACM, New York, APPENDIX Proof of Lemma 9. Let us first consier the case where r i {, +}, each value having probability /2. If a 2 b 2 then a c an the lemma hols with equality. Otherwise, let us write E T + cr + cr 2 2k E T + ar + br 2 2k S k 4 where S k T + 2c 2k + 2T 2k + T 2c 2k T + a + b 2k T + a b 2k T a + b 2k T a b 2k. We will show that S k 0 for all k 0. Since a 2 b 2 we can use the binomial theorem to expan every term other than 2T 2k in S k an get 2k S k 2T 2k 2k + T 2k i D i, i where i0 D i 2c i + 2c i a+b i a b i a+b i a b i. Observe now that for o i, D i 0. Moreover, we claim that D 2j 0 for all j. To see this claim observe that 2a 2 + 2b 2 a + b 2 + a b 2 an that for all j an x, y 0, x + y j x j + y j. Thus, S k 2T 2k + k j0 2k 2j k 2k T 2k j D 2j 2j j 0. T 2k j D 2j The proof for the case where r i { 3, 0, + 3} is merely a more cumbersome version of the proof above, so we omit it. That proof, though, brings forwar an interesting point. If one tries to take r i 0 with probability greater than 2/3, while maintaining a range of size 3 an variance, the lemma fails. In other wors, 2/3 is tight in terms of how much probability mass we can put to r i 0 an still have the all-ones vector be the worst-case one. 2] S. Vempala. A ranom sampling base algorithm for learning the intersection of half-spaces. In 38th Annual Symposium on Founations of Computer Science Miami, FL, 997, pages IEEE Comput. Soc. Press, Los Alamitos, CA, ] S. Vempala an R. I. Arriaga. An algorithmic theory of learning: robust concepts an ranom projection. In 40th Annual Symposium on Founations of Computer Science New York, NY, 999, pages IEEE Comput. Soc. Press, Los Alamitos, CA, 999.

Lecture Introduction. 2 Examples of Measure Concentration. 3 The Johnson-Lindenstrauss Lemma. CS-621 Theory Gems November 28, 2012

Lecture Introduction. 2 Examples of Measure Concentration. 3 The Johnson-Lindenstrauss Lemma. CS-621 Theory Gems November 28, 2012 CS-6 Theory Gems November 8, 0 Lecture Lecturer: Alesaner Mąry Scribes: Alhussein Fawzi, Dorina Thanou Introuction Toay, we will briefly iscuss an important technique in probability theory measure concentration

More information

Lower bounds on Locality Sensitive Hashing

Lower bounds on Locality Sensitive Hashing Lower bouns on Locality Sensitive Hashing Rajeev Motwani Assaf Naor Rina Panigrahy Abstract Given a metric space (X, X ), c 1, r > 0, an p, q [0, 1], a istribution over mappings H : X N is calle a (r,

More information

Acute sets in Euclidean spaces

Acute sets in Euclidean spaces Acute sets in Eucliean spaces Viktor Harangi April, 011 Abstract A finite set H in R is calle an acute set if any angle etermine by three points of H is acute. We examine the maximal carinality α() of

More information

Multi-View Clustering via Canonical Correlation Analysis

Multi-View Clustering via Canonical Correlation Analysis Technical Report TTI-TR-2008-5 Multi-View Clustering via Canonical Correlation Analysis Kamalika Chauhuri UC San Diego Sham M. Kakae Toyota Technological Institute at Chicago ABSTRACT Clustering ata in

More information

Least-Squares Regression on Sparse Spaces

Least-Squares Regression on Sparse Spaces Least-Squares Regression on Sparse Spaces Yuri Grinberg, Mahi Milani Far, Joelle Pineau School of Computer Science McGill University Montreal, Canaa {ygrinb,mmilan1,jpineau}@cs.mcgill.ca 1 Introuction

More information

Lecture 6 : Dimensionality Reduction

Lecture 6 : Dimensionality Reduction CPS290: Algorithmic Founations of Data Science February 3, 207 Lecture 6 : Dimensionality Reuction Lecturer: Kamesh Munagala Scribe: Kamesh Munagala In this lecture, we will consier the roblem of maing

More information

An Optimal Algorithm for Bandit and Zero-Order Convex Optimization with Two-Point Feedback

An Optimal Algorithm for Bandit and Zero-Order Convex Optimization with Two-Point Feedback Journal of Machine Learning Research 8 07) - Submitte /6; Publishe 5/7 An Optimal Algorithm for Banit an Zero-Orer Convex Optimization with wo-point Feeback Oha Shamir Department of Computer Science an

More information

Survey Sampling. 1 Design-based Inference. Kosuke Imai Department of Politics, Princeton University. February 19, 2013

Survey Sampling. 1 Design-based Inference. Kosuke Imai Department of Politics, Princeton University. February 19, 2013 Survey Sampling Kosuke Imai Department of Politics, Princeton University February 19, 2013 Survey sampling is one of the most commonly use ata collection methos for social scientists. We begin by escribing

More information

u!i = a T u = 0. Then S satisfies

u!i = a T u = 0. Then S satisfies Deterministic Conitions for Subspace Ientifiability from Incomplete Sampling Daniel L Pimentel-Alarcón, Nigel Boston, Robert D Nowak University of Wisconsin-Maison Abstract Consier an r-imensional subspace

More information

Linear First-Order Equations

Linear First-Order Equations 5 Linear First-Orer Equations Linear first-orer ifferential equations make up another important class of ifferential equations that commonly arise in applications an are relatively easy to solve (in theory)

More information

Topic 7: Convergence of Random Variables

Topic 7: Convergence of Random Variables Topic 7: Convergence of Ranom Variables Course 003, 2016 Page 0 The Inference Problem So far, our starting point has been a given probability space (S, F, P). We now look at how to generate information

More information

Computing Exact Confidence Coefficients of Simultaneous Confidence Intervals for Multinomial Proportions and their Functions

Computing Exact Confidence Coefficients of Simultaneous Confidence Intervals for Multinomial Proportions and their Functions Working Paper 2013:5 Department of Statistics Computing Exact Confience Coefficients of Simultaneous Confience Intervals for Multinomial Proportions an their Functions Shaobo Jin Working Paper 2013:5

More information

LECTURE NOTES ON DVORETZKY S THEOREM

LECTURE NOTES ON DVORETZKY S THEOREM LECTURE NOTES ON DVORETZKY S THEOREM STEVEN HEILMAN Abstract. We present the first half of the paper [S]. In particular, the results below, unless otherwise state, shoul be attribute to G. Schechtman.

More information

A Sketch of Menshikov s Theorem

A Sketch of Menshikov s Theorem A Sketch of Menshikov s Theorem Thomas Bao March 14, 2010 Abstract Let Λ be an infinite, locally finite oriente multi-graph with C Λ finite an strongly connecte, an let p

More information

Separation of Variables

Separation of Variables Physics 342 Lecture 1 Separation of Variables Lecture 1 Physics 342 Quantum Mechanics I Monay, January 25th, 2010 There are three basic mathematical tools we nee, an then we can begin working on the physical

More information

Lower Bounds for the Smoothed Number of Pareto optimal Solutions

Lower Bounds for the Smoothed Number of Pareto optimal Solutions Lower Bouns for the Smoothe Number of Pareto optimal Solutions Tobias Brunsch an Heiko Röglin Department of Computer Science, University of Bonn, Germany brunsch@cs.uni-bonn.e, heiko@roeglin.org Abstract.

More information

FLUCTUATIONS IN THE NUMBER OF POINTS ON SMOOTH PLANE CURVES OVER FINITE FIELDS. 1. Introduction

FLUCTUATIONS IN THE NUMBER OF POINTS ON SMOOTH PLANE CURVES OVER FINITE FIELDS. 1. Introduction FLUCTUATIONS IN THE NUMBER OF POINTS ON SMOOTH PLANE CURVES OVER FINITE FIELDS ALINA BUCUR, CHANTAL DAVID, BROOKE FEIGON, MATILDE LALÍN 1 Introuction In this note, we stuy the fluctuations in the number

More information

Multi-View Clustering via Canonical Correlation Analysis

Multi-View Clustering via Canonical Correlation Analysis Keywors: multi-view learning, clustering, canonical correlation analysis Abstract Clustering ata in high-imensions is believe to be a har problem in general. A number of efficient clustering algorithms

More information

Convergence of Random Walks

Convergence of Random Walks Chapter 16 Convergence of Ranom Walks This lecture examines the convergence of ranom walks to the Wiener process. This is very important both physically an statistically, an illustrates the utility of

More information

LATTICE-BASED D-OPTIMUM DESIGN FOR FOURIER REGRESSION

LATTICE-BASED D-OPTIMUM DESIGN FOR FOURIER REGRESSION The Annals of Statistics 1997, Vol. 25, No. 6, 2313 2327 LATTICE-BASED D-OPTIMUM DESIGN FOR FOURIER REGRESSION By Eva Riccomagno, 1 Rainer Schwabe 2 an Henry P. Wynn 1 University of Warwick, Technische

More information

A PAC-Bayesian Approach to Spectrally-Normalized Margin Bounds for Neural Networks

A PAC-Bayesian Approach to Spectrally-Normalized Margin Bounds for Neural Networks A PAC-Bayesian Approach to Spectrally-Normalize Margin Bouns for Neural Networks Behnam Neyshabur, Srinah Bhojanapalli, Davi McAllester, Nathan Srebro Toyota Technological Institute at Chicago {bneyshabur,

More information

16 Embeddings of the Euclidean metric

16 Embeddings of the Euclidean metric 16 Embeddings of the Euclidean metric In today s lecture, we will consider how well we can embed n points in the Euclidean metric (l 2 ) into other l p metrics. More formally, we ask the following question.

More information

Math 1B, lecture 8: Integration by parts

Math 1B, lecture 8: Integration by parts Math B, lecture 8: Integration by parts Nathan Pflueger 23 September 2 Introuction Integration by parts, similarly to integration by substitution, reverses a well-known technique of ifferentiation an explores

More information

Ramsey numbers of some bipartite graphs versus complete graphs

Ramsey numbers of some bipartite graphs versus complete graphs Ramsey numbers of some bipartite graphs versus complete graphs Tao Jiang, Michael Salerno Miami University, Oxfor, OH 45056, USA Abstract. The Ramsey number r(h, K n ) is the smallest positive integer

More information

Tractability results for weighted Banach spaces of smooth functions

Tractability results for weighted Banach spaces of smooth functions Tractability results for weighte Banach spaces of smooth functions Markus Weimar Mathematisches Institut, Universität Jena Ernst-Abbe-Platz 2, 07740 Jena, Germany email: markus.weimar@uni-jena.e March

More information

Agmon Kolmogorov Inequalities on l 2 (Z d )

Agmon Kolmogorov Inequalities on l 2 (Z d ) Journal of Mathematics Research; Vol. 6, No. ; 04 ISSN 96-9795 E-ISSN 96-9809 Publishe by Canaian Center of Science an Eucation Agmon Kolmogorov Inequalities on l (Z ) Arman Sahovic Mathematics Department,

More information

The derivative of a function f(x) is another function, defined in terms of a limiting expression: f(x + δx) f(x)

The derivative of a function f(x) is another function, defined in terms of a limiting expression: f(x + δx) f(x) Y. D. Chong (2016) MH2801: Complex Methos for the Sciences 1. Derivatives The erivative of a function f(x) is another function, efine in terms of a limiting expression: f (x) f (x) lim x δx 0 f(x + δx)

More information

A new proof of the sharpness of the phase transition for Bernoulli percolation on Z d

A new proof of the sharpness of the phase transition for Bernoulli percolation on Z d A new proof of the sharpness of the phase transition for Bernoulli percolation on Z Hugo Duminil-Copin an Vincent Tassion October 8, 205 Abstract We provie a new proof of the sharpness of the phase transition

More information

The Exact Form and General Integrating Factors

The Exact Form and General Integrating Factors 7 The Exact Form an General Integrating Factors In the previous chapters, we ve seen how separable an linear ifferential equations can be solve using methos for converting them to forms that can be easily

More information

Discrete Mathematics

Discrete Mathematics Discrete Mathematics 309 (009) 86 869 Contents lists available at ScienceDirect Discrete Mathematics journal homepage: wwwelseviercom/locate/isc Profile vectors in the lattice of subspaces Dániel Gerbner

More information

Time-of-Arrival Estimation in Non-Line-Of-Sight Environments

Time-of-Arrival Estimation in Non-Line-Of-Sight Environments 2 Conference on Information Sciences an Systems, The Johns Hopkins University, March 2, 2 Time-of-Arrival Estimation in Non-Line-Of-Sight Environments Sinan Gezici, Hisashi Kobayashi an H. Vincent Poor

More information

Quantum Mechanics in Three Dimensions

Quantum Mechanics in Three Dimensions Physics 342 Lecture 20 Quantum Mechanics in Three Dimensions Lecture 20 Physics 342 Quantum Mechanics I Monay, March 24th, 2008 We begin our spherical solutions with the simplest possible case zero potential.

More information

A Course in Machine Learning

A Course in Machine Learning A Course in Machine Learning Hal Daumé III 12 EFFICIENT LEARNING So far, our focus has been on moels of learning an basic algorithms for those moels. We have not place much emphasis on how to learn quickly.

More information

The chromatic number of graph powers

The chromatic number of graph powers Combinatorics, Probability an Computing (19XX) 00, 000 000. c 19XX Cambrige University Press Printe in the Unite Kingom The chromatic number of graph powers N O G A A L O N 1 an B O J A N M O H A R 1 Department

More information

Function Spaces. 1 Hilbert Spaces

Function Spaces. 1 Hilbert Spaces Function Spaces A function space is a set of functions F that has some structure. Often a nonparametric regression function or classifier is chosen to lie in some function space, where the assume structure

More information

Sharp Thresholds. Zachary Hamaker. March 15, 2010

Sharp Thresholds. Zachary Hamaker. March 15, 2010 Sharp Threshols Zachary Hamaker March 15, 2010 Abstract The Kolmogorov Zero-One law states that for tail events on infinite-imensional probability spaces, the probability must be either zero or one. Behavior

More information

. Using a multinomial model gives us the following equation for P d. , with respect to same length term sequences.

. Using a multinomial model gives us the following equation for P d. , with respect to same length term sequences. S 63 Lecture 8 2/2/26 Lecturer Lillian Lee Scribes Peter Babinski, Davi Lin Basic Language Moeling Approach I. Special ase of LM-base Approach a. Recap of Formulas an Terms b. Fixing θ? c. About that Multinomial

More information

Multi-View Clustering via Canonical Correlation Analysis

Multi-View Clustering via Canonical Correlation Analysis Kamalika Chauhuri ITA, UC San Diego, 9500 Gilman Drive, La Jolla, CA Sham M. Kakae Karen Livescu Karthik Sriharan Toyota Technological Institute at Chicago, 6045 S. Kenwoo Ave., Chicago, IL kamalika@soe.ucs.eu

More information

ensembles When working with density operators, we can use this connection to define a generalized Bloch vector: v x Tr x, v y Tr y

ensembles When working with density operators, we can use this connection to define a generalized Bloch vector: v x Tr x, v y Tr y Ph195a lecture notes, 1/3/01 Density operators for spin- 1 ensembles So far in our iscussion of spin- 1 systems, we have restricte our attention to the case of pure states an Hamiltonian evolution. Toay

More information

Lecture 6: Calculus. In Song Kim. September 7, 2011

Lecture 6: Calculus. In Song Kim. September 7, 2011 Lecture 6: Calculus In Song Kim September 7, 20 Introuction to Differential Calculus In our previous lecture we came up with several ways to analyze functions. We saw previously that the slope of a linear

More information

Generalizing Kronecker Graphs in order to Model Searchable Networks

Generalizing Kronecker Graphs in order to Model Searchable Networks Generalizing Kronecker Graphs in orer to Moel Searchable Networks Elizabeth Boine, Babak Hassibi, Aam Wierman California Institute of Technology Pasaena, CA 925 Email: {eaboine, hassibi, aamw}@caltecheu

More information

Qubit channels that achieve capacity with two states

Qubit channels that achieve capacity with two states Qubit channels that achieve capacity with two states Dominic W. Berry Department of Physics, The University of Queenslan, Brisbane, Queenslan 4072, Australia Receive 22 December 2004; publishe 22 March

More information

7.1 Support Vector Machine

7.1 Support Vector Machine 67577 Intro. to Machine Learning Fall semester, 006/7 Lecture 7: Support Vector Machines an Kernel Functions II Lecturer: Amnon Shashua Scribe: Amnon Shashua 7. Support Vector Machine We return now to

More information

Multi-View Clustering via Canonical Correlation Analysis

Multi-View Clustering via Canonical Correlation Analysis Kamalika Chauhuri ITA, UC San Diego, 9500 Gilman Drive, La Jolla, CA Sham M. Kakae Karen Livescu Karthik Sriharan Toyota Technological Institute at Chicago, 6045 S. Kenwoo Ave., Chicago, IL kamalika@soe.ucs.eu

More information

Schrödinger s equation.

Schrödinger s equation. Physics 342 Lecture 5 Schröinger s Equation Lecture 5 Physics 342 Quantum Mechanics I Wenesay, February 3r, 2010 Toay we iscuss Schröinger s equation an show that it supports the basic interpretation of

More information

Some Examples. Uniform motion. Poisson processes on the real line

Some Examples. Uniform motion. Poisson processes on the real line Some Examples Our immeiate goal is to see some examples of Lévy processes, an/or infinitely-ivisible laws on. Uniform motion Choose an fix a nonranom an efine X := for all (1) Then, {X } is a [nonranom]

More information

Robust Forward Algorithms via PAC-Bayes and Laplace Distributions. ω Q. Pr (y(ω x) < 0) = Pr A k

Robust Forward Algorithms via PAC-Bayes and Laplace Distributions. ω Q. Pr (y(ω x) < 0) = Pr A k A Proof of Lemma 2 B Proof of Lemma 3 Proof: Since the support of LL istributions is R, two such istributions are equivalent absolutely continuous with respect to each other an the ivergence is well-efine

More information

A Randomized Approximate Nearest Neighbors Algorithm - a short version

A Randomized Approximate Nearest Neighbors Algorithm - a short version We present a ranomize algorithm for the approximate nearest neighbor problem in - imensional Eucliean space. Given N points {x } in R, the algorithm attempts to fin k nearest neighbors for each of x, where

More information

DIFFERENTIAL GEOMETRY, LECTURE 15, JULY 10

DIFFERENTIAL GEOMETRY, LECTURE 15, JULY 10 DIFFERENTIAL GEOMETRY, LECTURE 15, JULY 10 5. Levi-Civita connection From now on we are intereste in connections on the tangent bunle T X of a Riemanninam manifol (X, g). Out main result will be a construction

More information

A Unified Theorem on SDP Rank Reduction

A Unified Theorem on SDP Rank Reduction A Unifie heorem on SDP Ran Reuction Anthony Man Cho So, Yinyu Ye, Jiawei Zhang November 9, 006 Abstract We consier the problem of fining a low ran approximate solution to a system of linear equations in

More information

Math 1271 Solutions for Fall 2005 Final Exam

Math 1271 Solutions for Fall 2005 Final Exam Math 7 Solutions for Fall 5 Final Eam ) Since the equation + y = e y cannot be rearrange algebraically in orer to write y as an eplicit function of, we must instea ifferentiate this relation implicitly

More information

Lecture 2 Lagrangian formulation of classical mechanics Mechanics

Lecture 2 Lagrangian formulation of classical mechanics Mechanics Lecture Lagrangian formulation of classical mechanics 70.00 Mechanics Principle of stationary action MATH-GA To specify a motion uniquely in classical mechanics, it suffices to give, at some time t 0,

More information

Lecture 5. Symmetric Shearer s Lemma

Lecture 5. Symmetric Shearer s Lemma Stanfor University Spring 208 Math 233: Non-constructive methos in combinatorics Instructor: Jan Vonrák Lecture ate: January 23, 208 Original scribe: Erik Bates Lecture 5 Symmetric Shearer s Lemma Here

More information

Table of Common Derivatives By David Abraham

Table of Common Derivatives By David Abraham Prouct an Quotient Rules: Table of Common Derivatives By Davi Abraham [ f ( g( ] = [ f ( ] g( + f ( [ g( ] f ( = g( [ f ( ] g( g( f ( [ g( ] Trigonometric Functions: sin( = cos( cos( = sin( tan( = sec

More information

Approximate Constraint Satisfaction Requires Large LP Relaxations

Approximate Constraint Satisfaction Requires Large LP Relaxations Approximate Constraint Satisfaction Requires Large LP Relaxations oah Fleming April 19, 2018 Linear programming is a very powerful tool for attacking optimization problems. Techniques such as the ellipsoi

More information

Lower Bounds for Local Monotonicity Reconstruction from Transitive-Closure Spanners

Lower Bounds for Local Monotonicity Reconstruction from Transitive-Closure Spanners Lower Bouns for Local Monotonicity Reconstruction from Transitive-Closure Spanners Arnab Bhattacharyya Elena Grigorescu Mahav Jha Kyomin Jung Sofya Raskhonikova Davi P. Wooruff Abstract Given a irecte

More information

Robust Low Rank Kernel Embeddings of Multivariate Distributions

Robust Low Rank Kernel Embeddings of Multivariate Distributions Robust Low Rank Kernel Embeings of Multivariate Distributions Le Song, Bo Dai College of Computing, Georgia Institute of Technology lsong@cc.gatech.eu, boai@gatech.eu Abstract Kernel embeing of istributions

More information

19 Eigenvalues, Eigenvectors, Ordinary Differential Equations, and Control

19 Eigenvalues, Eigenvectors, Ordinary Differential Equations, and Control 19 Eigenvalues, Eigenvectors, Orinary Differential Equations, an Control This section introuces eigenvalues an eigenvectors of a matrix, an iscusses the role of the eigenvalues in etermining the behavior

More information

Quantum mechanical approaches to the virial

Quantum mechanical approaches to the virial Quantum mechanical approaches to the virial S.LeBohec Department of Physics an Astronomy, University of Utah, Salt Lae City, UT 84112, USA Date: June 30 th 2015 In this note, we approach the virial from

More information

Lectures - Week 10 Introduction to Ordinary Differential Equations (ODES) First Order Linear ODEs

Lectures - Week 10 Introduction to Ordinary Differential Equations (ODES) First Order Linear ODEs Lectures - Week 10 Introuction to Orinary Differential Equations (ODES) First Orer Linear ODEs When stuying ODEs we are consiering functions of one inepenent variable, e.g., f(x), where x is the inepenent

More information

Math 342 Partial Differential Equations «Viktor Grigoryan

Math 342 Partial Differential Equations «Viktor Grigoryan Math 342 Partial Differential Equations «Viktor Grigoryan 6 Wave equation: solution In this lecture we will solve the wave equation on the entire real line x R. This correspons to a string of infinite

More information

6 General properties of an autonomous system of two first order ODE

6 General properties of an autonomous system of two first order ODE 6 General properties of an autonomous system of two first orer ODE Here we embark on stuying the autonomous system of two first orer ifferential equations of the form ẋ 1 = f 1 (, x 2 ), ẋ 2 = f 2 (, x

More information

SYNCHRONOUS SEQUENTIAL CIRCUITS

SYNCHRONOUS SEQUENTIAL CIRCUITS CHAPTER SYNCHRONOUS SEUENTIAL CIRCUITS Registers an counters, two very common synchronous sequential circuits, are introuce in this chapter. Register is a igital circuit for storing information. Contents

More information

Thermal conductivity of graded composites: Numerical simulations and an effective medium approximation

Thermal conductivity of graded composites: Numerical simulations and an effective medium approximation JOURNAL OF MATERIALS SCIENCE 34 (999)5497 5503 Thermal conuctivity of grae composites: Numerical simulations an an effective meium approximation P. M. HUI Department of Physics, The Chinese University

More information

Iterated Point-Line Configurations Grow Doubly-Exponentially

Iterated Point-Line Configurations Grow Doubly-Exponentially Iterate Point-Line Configurations Grow Doubly-Exponentially Joshua Cooper an Mark Walters July 9, 008 Abstract Begin with a set of four points in the real plane in general position. A to this collection

More information

TMA 4195 Matematisk modellering Exam Tuesday December 16, :00 13:00 Problems and solution with additional comments

TMA 4195 Matematisk modellering Exam Tuesday December 16, :00 13:00 Problems and solution with additional comments Problem F U L W D g m 3 2 s 2 0 0 0 0 2 kg 0 0 0 0 0 0 Table : Dimension matrix TMA 495 Matematisk moellering Exam Tuesay December 6, 2008 09:00 3:00 Problems an solution with aitional comments The necessary

More information

Math Notes on differentials, the Chain Rule, gradients, directional derivative, and normal vectors

Math Notes on differentials, the Chain Rule, gradients, directional derivative, and normal vectors Math 18.02 Notes on ifferentials, the Chain Rule, graients, irectional erivative, an normal vectors Tangent plane an linear approximation We efine the partial erivatives of f( xy, ) as follows: f f( x+

More information

Sturm-Liouville Theory

Sturm-Liouville Theory LECTURE 5 Sturm-Liouville Theory In the three preceing lectures I emonstrate the utility of Fourier series in solving PDE/BVPs. As we ll now see, Fourier series are just the tip of the iceberg of the theory

More information

EVALUATING HIGHER DERIVATIVE TENSORS BY FORWARD PROPAGATION OF UNIVARIATE TAYLOR SERIES

EVALUATING HIGHER DERIVATIVE TENSORS BY FORWARD PROPAGATION OF UNIVARIATE TAYLOR SERIES MATHEMATICS OF COMPUTATION Volume 69, Number 231, Pages 1117 1130 S 0025-5718(00)01120-0 Article electronically publishe on February 17, 2000 EVALUATING HIGHER DERIVATIVE TENSORS BY FORWARD PROPAGATION

More information

On combinatorial approaches to compressed sensing

On combinatorial approaches to compressed sensing On combinatorial approaches to compresse sensing Abolreza Abolhosseini Moghaam an Hayer Raha Department of Electrical an Computer Engineering, Michigan State University, East Lansing, MI, U.S. Emails:{abolhos,raha}@msu.eu

More information

Analyzing Tensor Power Method Dynamics in Overcomplete Regime

Analyzing Tensor Power Method Dynamics in Overcomplete Regime Journal of Machine Learning Research 18 (2017) 1-40 Submitte 9/15; Revise 11/16; Publishe 4/17 Analyzing Tensor Power Metho Dynamics in Overcomplete Regime Animashree Ananumar Department of Electrical

More information

Free rotation of a rigid body 1 D. E. Soper 2 University of Oregon Physics 611, Theoretical Mechanics 5 November 2012

Free rotation of a rigid body 1 D. E. Soper 2 University of Oregon Physics 611, Theoretical Mechanics 5 November 2012 Free rotation of a rigi boy 1 D. E. Soper 2 University of Oregon Physics 611, Theoretical Mechanics 5 November 2012 1 Introuction In this section, we escribe the motion of a rigi boy that is free to rotate

More information

Leaving Randomness to Nature: d-dimensional Product Codes through the lens of Generalized-LDPC codes

Leaving Randomness to Nature: d-dimensional Product Codes through the lens of Generalized-LDPC codes Leaving Ranomness to Nature: -Dimensional Prouct Coes through the lens of Generalize-LDPC coes Tavor Baharav, Kannan Ramchanran Dept. of Electrical Engineering an Computer Sciences, U.C. Berkeley {tavorb,

More information

PDE Notes, Lecture #11

PDE Notes, Lecture #11 PDE Notes, Lecture # from Professor Jalal Shatah s Lectures Febuary 9th, 2009 Sobolev Spaces Recall that for u L loc we can efine the weak erivative Du by Du, φ := udφ φ C0 If v L loc such that Du, φ =

More information

Implicit Differentiation

Implicit Differentiation Implicit Differentiation Thus far, the functions we have been concerne with have been efine explicitly. A function is efine explicitly if the output is given irectly in terms of the input. For instance,

More information

Entanglement is not very useful for estimating multiple phases

Entanglement is not very useful for estimating multiple phases PHYSICAL REVIEW A 70, 032310 (2004) Entanglement is not very useful for estimating multiple phases Manuel A. Ballester* Department of Mathematics, University of Utrecht, Box 80010, 3508 TA Utrecht, The

More information

Counting Lattice Points in Polytopes: The Ehrhart Theory

Counting Lattice Points in Polytopes: The Ehrhart Theory 3 Counting Lattice Points in Polytopes: The Ehrhart Theory Ubi materia, ibi geometria. Johannes Kepler (1571 1630) Given the profusion of examples that gave rise to the polynomial behavior of the integer-point

More information

Chapter 4. Electrostatics of Macroscopic Media

Chapter 4. Electrostatics of Macroscopic Media Chapter 4. Electrostatics of Macroscopic Meia 4.1 Multipole Expansion Approximate potentials at large istances 3 x' x' (x') x x' x x Fig 4.1 We consier the potential in the far-fiel region (see Fig. 4.1

More information

arxiv: v2 [cs.ds] 11 May 2016

arxiv: v2 [cs.ds] 11 May 2016 Optimizing Star-Convex Functions Jasper C.H. Lee Paul Valiant arxiv:5.04466v2 [cs.ds] May 206 Department of Computer Science Brown University {jasperchlee,paul_valiant}@brown.eu May 3, 206 Abstract We

More information

arxiv: v4 [cs.ds] 7 Mar 2014

arxiv: v4 [cs.ds] 7 Mar 2014 Analysis of Agglomerative Clustering Marcel R. Ackermann Johannes Blömer Daniel Kuntze Christian Sohler arxiv:101.697v [cs.ds] 7 Mar 01 Abstract The iameter k-clustering problem is the problem of partitioning

More information

The Three-dimensional Schödinger Equation

The Three-dimensional Schödinger Equation The Three-imensional Schöinger Equation R. L. Herman November 7, 016 Schröinger Equation in Spherical Coorinates We seek to solve the Schröinger equation with spherical symmetry using the metho of separation

More information

A note on asymptotic formulae for one-dimensional network flow problems Carlos F. Daganzo and Karen R. Smilowitz

A note on asymptotic formulae for one-dimensional network flow problems Carlos F. Daganzo and Karen R. Smilowitz A note on asymptotic formulae for one-imensional network flow problems Carlos F. Daganzo an Karen R. Smilowitz (to appear in Annals of Operations Research) Abstract This note evelops asymptotic formulae

More information

Necessary and Sufficient Conditions for Sketched Subspace Clustering

Necessary and Sufficient Conditions for Sketched Subspace Clustering Necessary an Sufficient Conitions for Sketche Subspace Clustering Daniel Pimentel-Alarcón, Laura Balzano 2, Robert Nowak University of Wisconsin-Maison, 2 University of Michigan-Ann Arbor Abstract This

More information

Lagrangian and Hamiltonian Mechanics

Lagrangian and Hamiltonian Mechanics Lagrangian an Hamiltonian Mechanics.G. Simpson, Ph.. epartment of Physical Sciences an Engineering Prince George s Community College ecember 5, 007 Introuction In this course we have been stuying classical

More information

1 dx. where is a large constant, i.e., 1, (7.6) and Px is of the order of unity. Indeed, if px is given by (7.5), the inequality (7.

1 dx. where is a large constant, i.e., 1, (7.6) and Px is of the order of unity. Indeed, if px is given by (7.5), the inequality (7. Lectures Nine an Ten The WKB Approximation The WKB metho is a powerful tool to obtain solutions for many physical problems It is generally applicable to problems of wave propagation in which the frequency

More information

On the Surprising Behavior of Distance Metrics in High Dimensional Space

On the Surprising Behavior of Distance Metrics in High Dimensional Space On the Surprising Behavior of Distance Metrics in High Dimensional Space Charu C. Aggarwal, Alexaner Hinneburg 2, an Daniel A. Keim 2 IBM T. J. Watson Research Center Yortown Heights, NY 0598, USA. charu@watson.ibm.com

More information

Chromatic number for a generalization of Cartesian product graphs

Chromatic number for a generalization of Cartesian product graphs Chromatic number for a generalization of Cartesian prouct graphs Daniel Král Douglas B. West Abstract Let G be a class of graphs. The -fol gri over G, enote G, is the family of graphs obtaine from -imensional

More information

On colour-blind distinguishing colour pallets in regular graphs

On colour-blind distinguishing colour pallets in regular graphs J Comb Optim (2014 28:348 357 DOI 10.1007/s10878-012-9556-x On colour-blin istinguishing colour pallets in regular graphs Jakub Przybyło Publishe online: 25 October 2012 The Author(s 2012. This article

More information

Parameter estimation: A new approach to weighting a priori information

Parameter estimation: A new approach to weighting a priori information Parameter estimation: A new approach to weighting a priori information J.L. Mea Department of Mathematics, Boise State University, Boise, ID 83725-555 E-mail: jmea@boisestate.eu Abstract. We propose a

More information

arxiv:physics/ v4 [physics.class-ph] 9 Jul 1999

arxiv:physics/ v4 [physics.class-ph] 9 Jul 1999 AIAA-99-2144 PROPULSION THROUGH ELECTROMAGNETIC SELF-SUSTAINED ACCELERATION arxiv:physics/9906059v4 [physics.class-ph] 9 Jul 1999 Abstract As is known the repulsion of the volume elements of an uniformly

More information

Proof of SPNs as Mixture of Trees

Proof of SPNs as Mixture of Trees A Proof of SPNs as Mixture of Trees Theorem 1. If T is an inuce SPN from a complete an ecomposable SPN S, then T is a tree that is complete an ecomposable. Proof. Argue by contraiction that T is not a

More information

Vectors in two dimensions

Vectors in two dimensions Vectors in two imensions Until now, we have been working in one imension only The main reason for this is to become familiar with the main physical ieas like Newton s secon law, without the aitional complication

More information

arxiv: v4 [math.pr] 27 Jul 2016

arxiv: v4 [math.pr] 27 Jul 2016 The Asymptotic Distribution of the Determinant of a Ranom Correlation Matrix arxiv:309768v4 mathpr] 7 Jul 06 AM Hanea a, & GF Nane b a Centre of xcellence for Biosecurity Risk Analysis, University of Melbourne,

More information

Two formulas for the Euler ϕ-function

Two formulas for the Euler ϕ-function Two formulas for the Euler ϕ-function Robert Frieman A multiplication formula for ϕ(n) The first formula we want to prove is the following: Theorem 1. If n 1 an n 2 are relatively prime positive integers,

More information

d-dimensional Arrangement Revisited

d-dimensional Arrangement Revisited -Dimensional Arrangement Revisite Daniel Rotter Jens Vygen Research Institute for Discrete Mathematics University of Bonn Revise version: April 5, 013 Abstract We revisit the -imensional arrangement problem

More information

05 The Continuum Limit and the Wave Equation

05 The Continuum Limit and the Wave Equation Utah State University DigitalCommons@USU Founations of Wave Phenomena Physics, Department of 1-1-2004 05 The Continuum Limit an the Wave Equation Charles G. Torre Department of Physics, Utah State University,

More information

Euler equations for multiple integrals

Euler equations for multiple integrals Euler equations for multiple integrals January 22, 2013 Contents 1 Reminer of multivariable calculus 2 1.1 Vector ifferentiation......................... 2 1.2 Matrix ifferentiation........................

More information

d dx But have you ever seen a derivation of these results? We ll prove the first result below. cos h 1

d dx But have you ever seen a derivation of these results? We ll prove the first result below. cos h 1 Lecture 5 Some ifferentiation rules Trigonometric functions (Relevant section from Stewart, Seventh Eition: Section 3.3) You all know that sin = cos cos = sin. () But have you ever seen a erivation of

More information

Exercise 1. Exercise 2.

Exercise 1. Exercise 2. Exercise. Magnitue Galaxy ID Ultraviolet Green Re Infrare A Infrare B 9707296462088.56 5.47 5.4 4.75 4.75 97086278435442.6.33 5.36 4.84 4.58 2255030735995063.64.8 5.88 5.48 5.4 56877420209795 9.52.6.54.08

More information

Lecture 18: March 15

Lecture 18: March 15 CS71 Randomness & Computation Spring 018 Instructor: Alistair Sinclair Lecture 18: March 15 Disclaimer: These notes have not been subjected to the usual scrutiny accorded to formal publications. They may

More information