Fast Spectral Clustering via the Nyström Method

Size: px
Start display at page:

Download "Fast Spectral Clustering via the Nyström Method"

Transcription

1 Fast Spectra Custering via the Nyström Method Anna Choromanska, Tony Jebara, Hyungtae Kim, Mahesh Mohan 3, and Caire Monteeoni 3 Department of Eectrica Engineering, Coumbia University, NY, USA Department of Computer Science, Coumbia University, NY, USA 3 Department of Computer Science, George Washington University, DC, USA {aec63, tj8, hk56}@coumbia.edu, {mahesh mohan, cmonte}@gwu.edu Abstract. We propose and anayze a fast spectra custering agorithm with computationa compexity inear in the number of data points that is directy appicabe to arge-scae datasets. The agorithm combines two powerfu techniques in machine earning: spectra custering agorithms and Nyström methods commony used to obtain good quaity ow rank approximations of arge matrices. The proposed agorithm appies the Nyström approximation to the graph Lapacian to perform custering. We provide theoretica anaysis of the performance of the agorithm and show the error bound it achieves and we discuss the conditions under which the agorithm performance is comparabe to spectra custering with the origina graph Lapacian. We aso present empirica resuts. Keywords: spectra custering, Nyström method, arge-scae custering, samping, sparsity, performance guarantees, error bounds, unsupervised earning Introduction Custering is one of the fundamenta probems in machine earning. The recent widespread deveopment of sensors, data-storage and data-acquisition devices has heped make arge data-sets common pace. This, however, poses a serious computationa chaenge for the existing custering techniques. Spectra custering techniques (Luxburg, 7) are widey used, due to their simpicity and empirica performance advantages compared to other custering methods, such as k-means or singe-inkage agorithms. However, a significant obstace to scaing up spectra custering to arge datasets is that it requires buiding an affinity matrix between pairs of data points which becomes computationay prohibitive for arge data-sets. There have been severa attempts to address this probem and make spectra custering agorithms more appicabe to arge-scae probems. Here we study an Main contact author: Anna Choromanska, e-mai: aec63@coumbia.edu, maiing adress: Department of Eectrica Engineering, Coumbia University, CEPSR 64, 4 Amsterdam Avenue, 7, NY, USA.

2 A. Choromanska et a. approach that extends the spectra custering agorithm, described in Ng et a. (), via Nyström approximation techniques. Our work is most reated to Wiiams and Seeger (); Fowkes et a. (4); Li et a. (), which use the Nyström method to sampe the coumns of the affinity matrix and further approximate the fu matrix by using correations between the samped coumns and the remaining coumns (Fowkes et a., 4). However, these works did not provide performance guarantees; that is our primary contribution. Other approaches to scaing up spectra custering incude work by Yan et a. (9), which used the k-means custering agorithm (Loyd, 98) as a preprocessing step to spectra custering, to reduce its computationa compexity. The anaysis assumes the data are generated by a mixture mode (the same assumption is made in the work by Lashkari and Goand (7)). Reated work by Drineas and Mahoney (5) performs non-uniform samping of the Gram matrix and provides a bound on the approximation error, however in order to achieve good performance one may need to sampe arge number of coumns (in specia cases even O(n)) and, furthermore, the practicaity of this technique for massive datasets may be imited (Yan et a., 9). Severa other works on constructing approximations that are tighter than the Nyström method s, for sparse graph Lapacians, have aso emerged (Fung et a., ; Spieman and Teng, ). However, the computationa compexity of these methods depends highy on the number of edges in the graph. In contrast, the Nyström method has a fixed compexity that is dependent ony on the number of vertices n and the number of samped coumns. This is potentiay more usefu in cases where a arge number of edges exist but ony a few have significanty arge weights, as is the case in many sparse datasets that arise in appications, such as coaborative fitering. This paper combines the spectra custering agorithm (Ng et a., ) with the Nyström approximation method by using a Nyström approximation to the graph Lapacian. Our anaysis differs from the approach of Bekin and Niyogi (7) in that we focus on the finite sampe anaysis whereas Bekin and Niyogi (7) emphasized asymptotic resuts. In particuar, they show that if points are samped uniformy at random from an unknown sub-manifod M R N, then the eigenvectors of a suitaby constructed graph Lapacian converge to the eigenfunctions of the Lapace-Betrami operator on M. Our approach eads to a practica agorithm with compexity inear in the number of data points n. We provide performance guarantees for this agorithm by combining Nyström approximation anaysis, using a uniform random samping without repacement scheme due to Kumar et a. (9), with perturbation theory anaysis (Ng et a., ). We discuss conditions under which the agorithm s performance is comparabe to spectra custering with the origina graph Lapacian. Approach. Spectra Custering Agorithm

3 Agorithm Spectra custering Fast Spectra Custering via the Nyström Method 3 Input: dataset S = {s, s..., s n } R d, number of custers k, kerne function κ : R d R d R Output: k-custering of S A R n n s.t. A ij = δ[i j]κ(s i, s j) D R n n s.t. D ij = δ[i = j] n j= Aij L = I D AD X R n k s.t. SmaestEigenVectors(L, k) Y R n k s.t. Y im = X im/ m X im K = CusterRows(Y) (via any k-custering agorithm minimizing distortion, i.e. k- means) In genera, spectra custering methods can be interpreted as graph partitioning agorithms and the above agorithm (Agorithm ) can be seen as graph partitioning with a normaized-cut cost function. Agorithm shows the widey used normaized spectra custering agorithm presented in Ng et a. (). Given the set of n points S = {s, s..., s n } the agorithm first buids an n n affinity matrix A, i.e.: A ij = κ(s i, s j ) if i j and otherwise. Here A ij corresponds to the i th row and j th coumn of the affinity matrix and κ is any kerne function accepting two input data-points and returning a scaar output. Once the affinity matrix is computed, the normaized graph Lapacian L can be constructed. The first k eigenvectors of L are then normaized and custered. It was shown in Ng et a. () that one can perform spectra k- custering using a perturbed version à of the idea affinity matrix A. Under certain assumptions, the custerings obtained using A and à wi be simiar. Our goa is to extend these assumptions and show that using cose to idea graph Lapacian L and its Nyström r-rank approximation, L wi aso give simiar custering resuts. Based on the anaysis in Ng et a. () we know that if the four assumptions isted beow are satisfied then using either à or A to perform spectra custering wi give simiar partitionings of the dataset (and aso simiar to the true custering of the dataset): Assumption A: γ> i={,,...,k} λ i γ, where λ i is the second argest eigenvaue of L i, where L i is the subbock of L corresponding to custer i. à Assumption A: ϵ> j i,i ={,,...,k},i i j S i S i ϵ d, where = m S i à jm and d = m S i à m and S i is the set of points beonging to the i th custer. : S à j Assumption A3: ϵ > i i={,,...,k},j Si ϵ d ( à m j ) d. dm Assumption A4: C> d i={,,...,k},j={,,...,ni} j ( n i d = )/(Cn i ). Assumption A guarantees each custer to be tight. Assumption A and A3 require data points within a custer to be more connected to each other than

4 4 A. Choromanska et a. they are with data points from any other custer. Finay, the ast assumption requires that the points in any custer can never be much ess connected than other points in the same custer. The simiarity of the custerings obtained using A and à is then assured via Theorem. Let yi j be the jth row of Y i from Agorithm, where Y i is the subbock of Y corresponding to custer i. Then the foowing theorem hods. Theorem (Ng et a. ()). Let assumptions A, A, A3 and A4 hod. Set ϵ = k(k )ϵ + kϵ. If γ > ( + )ϵ, then there exist k orthonorma vectors r, r,..., r k such that Y in Agorithm satisfies n k n i yj i r i 4C(4 + k) ϵ (γ ϵ). i= j=. Nyström Method for Matrix Approximation Agorithm Nyström method for matrix approximation : Input: matrix L, - number of coumns samped, r - rank approximation (r << n) : Output: Σ and Ũ such that L = Ũ ΣŨ 3: L indices of coumns samped 4: C G(:, L) 5: W C(L, :) 6: W r best r-rank approximation to W 7: Σ = n Σ W r and Ũ = n CU W r Σ W r, where W r = U Wr Σ Wr U W r We now expicate the Nyström r-rank approximation for any symmetric positive semidefinite (SPSD) matrix L R n n. After performing samping (we wi ony be using uniform samping without repacement schemes), create matrix C R n from the samped coumns. Then, form matrix W R matrix consisting of the intersection of these coumns with the corresponding rows of L. Let W = UΣU, where U is orthogona and Σ = diag(σ, σ,..., σ ) is a rea diagona matrix with the diagona sorted in decreasing order. Let W r + be the pseudo-inverse of the best rank-r approximation to W (W r + = r t= σ t U (t) U (t), where U (t) and U (t) are respectivey the t th coumn and row of U). Then the Nyström approximation L of L can be obtained as foows: L = CW + r C. Furthermore if we represent L as L = Ũ ΣŨ then Σ = n Σ W r and Ũ = n CU W r Σ W r, where W r = U Wr Σ Wr U W r. Theorem due to Kumar et a. (9) shows the performance bounds for the Nyström method when used with uniform samping without repacement. In Kumar et a. (9) the authors aso

5 Fast Spectra Custering via the Nyström Method 5 compare the quaity of obtained Nyström approximations, on the experiments with arge-scae datasets, when using uniform and non-uniform samping strategies (they consider both samping with and without repacement). They consider two most popuar non-uniform samping techniques: coumn-norm samping and diagona samping. They show that uniform samping without repacement is not ony more effcient both in time and space but aso improves the accuracy of the Nyström method. Theorem (Kumar et a. (9)). Let G R n n be an SPSD matrix. Assume that coumns of G are samped uniformy at random without repacement, et G r be the rank-r Nyström approximation to G and et G r be the best rank-r approximation to G. Let ϵ >, 64r/ϵ 4 og(/δ)ξ(,n ) and η =, where mu ξ(m, u) =. Then with probabiity at east δ, m+u / /( max{m,u}) G G r F G G r F + ϵ n i D() G ii n n G ii + η max(ng ii) i=, where F is the Frobenius norm, i D() G ii is the sum of the argest diagona entries of G. 3 Fast Spectra Custering Agorithm Agorithm 3 Fast spectra custering Input: dataset S = {s, s..., s n } R d, k - number of custers, - number of coumns samped, r - rank approximation (k r << n) Output: k-custering of S L indices of coumns samped (uniformy without repacement) Â A(:, L) D R n n s.t. D ij = δ[i = j]/ j= Âij n i= Âij R s.t. ij = δ[i = j]/ C Î D Â : I - matrix of coumns of I indexed by L n W C(L, :) W r best r-rank approximation to W Σ = n ΣW r and Ũ = n CUW r Σ W r, where W r = U Wr Σ Wr U W r X SmaestEigenVectors(Ũ, k) Y NormaizeRows(X ) : Y im = X im/( m X im) K CusterRows(Y ) (use any k-custering agorithm minimizing distortion, i.e. k- means)

6 6 A. Choromanska et a. For arge-scae fast spectra custering, we propose Agorithm 3. The agorithm chooses coumns samped uniformy at random from the affinity matrix. It therefore never buids the entire n n affinity matrix which woud be computationay prohibitive. It then computes two sparse diagona degree matrices D AND. Subsequenty, the matrix C is recovered which is n x. Matrix C pays the roe of the samped graph Lapacian. We then foow the steps of Agorithm to obtain the approximate eigensystem of the graph Lapacian and finay the first k eigenvectors are normaized and custered. Ceary, Agorithm 3 performs samping of the affinity matrix. This is in contrast to the more computationay expensive approach of computing the compete n n affinity matrix and then obtaining matrix C by samping directy from the graph Lapacian. We provide Theorem 3 to show that for appropriate vaues of, both of these agorithms wi give simiar custering resuts. First, et us introduce some additiona notation. We consider two scenarios: samping the graph Lapacian and samping the affinity matrix. Let L be the set of indices of samped coumns. Let I be the matrix of coumns of I that are indexed by L. Notice that for i {,,..., n} and j L, any entry in the samped graph Lapacian has the foowing form: C ij = I ij A ij ( n a= A aj)( n b= A ib). On the other hand, matrix C in Agorithm 3 has the foowing form: C ij = I A ij ij n ( n a= A aj)(. b L A ib) The difference ies in the second term in the denominator and the scaing factor. Consider Theorem 3. n Theorem 3. Let A ij s be iid scaar random variabes (bounded in [, ]) whose expectation is µ. With probabiity at east δ the foowing hods: C µ ij im µ + δ C ij C µ ij max(, ), n µ δ where δ = og(/δ). n Proof. By the aw of arge numbers we have that im n n b= A ib = µ. By Hoeffding s inequaity we have that with probabiity at east δ, b L A ib µ og(/δ). Therefore im C ij = im n n [ I ij n n (I ij C ij) n n b= A ib b L A ib The iid assumption is made ony for the purpose of this section. ]

7 Finay, if and if = im n b L A ib µ: C ij b L A ib < µ: Fast Spectra Custering via the Nyström Method 7 [ C ij µ µ + δ im n C ij im n [ C ij [ C ij n n b= A ib b L A ib ] n n b= A ib b L A ib n n b= A ib b L A ib ] C ij ] C ij, µ µ δ, where δ = og(/δ). Combining both cases gives the theorem. Theorem 3 shows that, for sufficienty arge, the two agorithms under consideration (Agorithm 3 samping the affinity matrix and the sower aternative of samping the graph Lapacian) shoud produce simiar C matrices and thus yied simiar custering resuts. Furthermore, in batch settings with finite n, Agorithm 3 is sti appicabe, i.e. consider the exampe presented on Figure showing the partitionings of two simpe datasets obtained by spectra custering agorithm of Ng et a. () using the fu affinity matrix and, for comparison, our Agorithm 3. The size of both datasets is very sma (n = 5), but the performance of both agorithms is very simiar. Finay, in our theoretica anaysis we wi focus on the scenario when the graph Lapacian is being samped. This anaysis is easier than considering samping the affinity matrix which, for instance, does not need to be PSD. We end this section with Theorem 4 showing the computationa compexity of the proposed Agorithm 3. a) b) Fig.. Resut of spectra custering on two datasets (a and b), n = 5, = % n, r =. Top row: the dataset (eft) and the partitioning obtained by spectra custering using the fu affinity matrix (right). Bottom row: the dataset with samped data points (green) (eft) and the partitioning obtained by Agorithm 3 (using the samped affinity matrix) (right).

8 8 A. Choromanska et a. Theorem 4. The computationa compexity of Agorithm 3 is O(n max(r, c)) + Γ, where c is the cost of evauating a singe kerne function between two data points and Γ is the cost of the custering agorithm minimizing distortion used to obtain the fina custering. 4 Performance Guarantees As was mentioned before the theoretica anaysis considers the case where we sampe the graph Lapacian buit from the n n affinity matrix and thus C is an n matrix of samped coumns. Furthermore, matrix W is an matrix consisting of the intersection of these coumns with the corresponding rows of the graph Lapacian. Our theoretica anaysis wi consider the performance of the proposed agorithm in the case where the affinity matrix and the corresponding graph Lapacian are cose to bock diagona matrices. In particuar we wi require that L L r F L L r F ϵn, where L is the idea graph Lapacian, L is the true, cose to diagona, graph Lapacian that is samped and L r is its best rank r Nyström approximation. Two more conditions that we assume hods wi be introduced ater. This section is organized as foows: we wi first show the main resut (Theorem 5) and then we wi show the technica emmas and proofs that ed to this resut. 4. Main Resut Let A be the idea affinity matrix that gave rise to the idea graph Lapacian L. Let L be the Nyström r-rank approximation to L and et à be the affinity matrix that woud give rise to L in case when no Nyström approximation was used. We wi now present our main resut, Theorem 5. Let yj i be the jth row of Y i from Agorithm 3, where Y i is the subbock of Y corresponding to custer i. Then the foowing theorem hods. Theorem 5. Let ϵ >, 64r/ϵ 4 og(/δ)ξ(,n ) and η =, where ξ(m, u) = mu m+u / /( max{m,u}). Let γ, ϵ, ϵ and C be defined as in Lemma,, 3 and 4. Set ϵ = k(k )ϵ + kϵ. If γ > ( + )ϵ, then with probabiity at east δ, there exist k orthogona vectors r, r,..., r k (ri r j = if i = j, otherwise) so that Y in Agorithm 3 satisfies: n k n i yj i r i 4C(4 + k) ϵ i= j= (γ ϵ ) Theorem 5 is generaization of Theorem. It differs from Theorem in that it extends the four assumptions used in Theorem which resut from the fact that à is a very specia version of the perturbed idea A, in particuar it is an affinity matrix that gave rise to the Nyström r-rank approximation to the graph

9 Fast Spectra Custering via the Nyström Method 9 Lapacian. The assumption on each λ i ensures that each custer is tight enough such that after samping the custers wi sti remain tight (γ can be interpreted as the measure of tigtness of each custer after samping). This assumption aso shows that when we decrease the number of samped coumns, we expect the origina custers to be tighter in order for the custers obtained after samping to aso be tight enough such that the dataset is sti k-custerabe. 4. Theoretica Anaysis We wi first present Theorem 6 which is a version of Theorem when the samped matrix is a graph Lapacian L. Theorem 6 reies on the fact that L is a SPSD matrix that is cose to bock diagona. Theorem 6. Let L R n n be an idea graph Lapacian and L be the cose to bock diagona graph Lapacian defined before. Assume that coumns of L are samped uniformy at random without repacement and et L be the best rank-r Nyström approximation to L. Let ϵ >, 64r/ϵ 4 and η = where ξ(m, u) = mu m+u / /( max{m,u}) L L F ϵn + η. og(/δ)ξ(,n ),. Then with probabiity at east δ, Reca a usefu theorem (Theorem 7) that we wi need ater. It can be found i.e. in Kannan and Vempaa (9). Intuitivey Theorem 7 impies that if two matrices are cose (in terms of the squared Frobenius norm of their difference), then their singuar vaues shoud aso be cose too. Theorem 7. For any two n n symmetric matrices A and B, n (σ t (A) σ t (B)) A B F t= We now proceed with the theoretica anaysis that wi ead to Theorem 5. We aim to make use of Theorem and then Theorem 6 to provide theoretica guarantees on the performance of spectra custering when using the Nyström approximation to the cose to idea ( cose to bock diagona) graph Lapacian. We wi focus on extending assumptions A, A, A3 and A4 used in Theorem. We wi present Lemma, 3, 4 and. Appying them to Theorem yieds our main resut captured in Theorem 5. We wi aso make three additiona assumptions that we wi assume hod throughout the entire anaysis. First of a we wi assume that ( λ i λ i ) r n t= ( λ i t λ i t). Secondy we wi assume that h n i ñ i h, where n i is the number of data points assigned to custer i when using L, ñ i is the number of data points assigned to custer i when using L rather than L and h is some positive constant. Finay since the origina Lapacian is

10 A. Choromanska et a. assumed to be cose to bock diagona, we woud ike its Nyström approximation to be cose to bock diagona by assuming that the foowing two conditions hods: j S i S i L j L j = n i S i L j L j f n L L, where i i, j = arg max j Si S i L j L j and f > is some constant, and j, S i L j L j g n L L, where g > is some constant. Lemma. Let λ i be the second argest eigenvaue of L i, where L i is the subbock of L corresponding to custer i, and et λ i be the second argest eigenvaue of L i, where L i is the subbock of L corresponding to custer i. If λ i cϵ r( + η) (c > is some constant) then with probabiity at east δ, γ> λi γ. Proof. We know that ( λ i λ i ) n r ( λ i t λ i t) n L i L i F n L L F, () t= where the first inequaity comes from Theorem 7. By appying Jensen s inequaity to the eft hand side of Equation, we obtain r λ i t λ i t r n L L F. () t= Then in particuar the foowing hods: λ i λ i r n L L F. (3) By assumption, we know that λ i cϵ r( + η). Now, if λ i λ i then emma hods. If λ i > λ i, then we can rewrite Equation 3 as: λ i λ i + r n L L F (4) Since λ i cϵ r( + η) and by Theorem 6 with probabiity at east δ the foowing hods: r n L L F ϵ r( + η), then we can write that with probabiity at east δ, γ> λi γ, where γ = (c )ϵ r( + η). Lemma extends assumption A from Ng et a. (). Before we proceed to the next emma, et us first introduce some more notation. We know that à is defined as the affinity matrix that woud give rise to graph Lapacian L in case when no Nyström approximation was used and thus L = I D / / à D (in this case D is the diagona matrix whose (i, i)-eement is the sum of à s i th row). Let j S i, where i {,,..., k}. Define the foowing: d (j) = n m= A jm, d(j) = n m= Ãjm, d j = m S i A jm, dj = m S i à jm. Notice that d (j) d j. Let d (i) i {,,...,k} = min j Si dj and d = min d (i) i {,,...,k}. Aso, et D (i) i {,,...,k} = max j Si d(j) and D = max D (i) i {,,...,k}. At

11 Fast Spectra Custering via the Nyström Method this point we wi make a reasonabe assumption that D d is a bounded positive constant. Assuming the dataset has baanced custers (i.e., no custer is significanty bigger/smaer than any other) and in particuar the datasets have no outiers, this assumption wi be naturay satisfied. Furthermore, et α Si S i = min j Si, S i,i,i {,,...,k} d d (j) d() and et α = min i,i {,,...,k} α Si S i. Note that α (, ] and in the idea case α =. We are now ready to state and prove Lemma 3. Lemma. C> i={,,...,k},j={,,...,ni } ( n i = d )/(Cn i ). Proof. Consider any i {,,..., k} and any j, S i. It is true that where C = D d ni d = d D n i Cn i, (5) is a bounded positive constant as was aready discussed before. Lemma extends assumption A4 from Ng et a. (). Lemma 3. With probabiity at east δ, i,i ={,,...,k},i i ϵ, where ϵ = ϵc f + η (f > is some constant). Proof. Let j = arg max j Si S i L j L j. We know that: j S i S i L j L j j S i j S i S i à j d S i L j L j f n L L (6) The eft-hand side of Equation 6 can be further expressed as j S i S i à j d(j) d() A j = d(j) d () Combining this resut with Equation 6 we have Rewrite Equation 8 as: j S i S i j S i S i à j d j S i S i à j d(j) d(). (7) à j d (j) d() f n L L F. (8) dj d d (j) d() f n L L F. (9) The eft-hand side of Equation 9 is ower-bounded by α j S i S i thus j S i S i à j à j d and f d α n L L F. ()

12 A. Choromanska et a. Again, by Theorem 6 we can write that with probabiity at east δ the foowing hods: ( ) à j f ϵn D + η ϵf + η = ϵc d n α d f + η, () j S i S i where the ast inequaity comes from the fact that α ( d D ). Lemma 3 extends assumption A from Ng et a. (). Notice that exacty the same proof technique coud be used to show that à j d ϵ j d n i. This resut and β = max i {,,...,k} β Si. We can now pro- wi be used in the next emma. d Define β Si = max j Si, S i ceed to the next emma. S i : S à j Lemma 4. With probabiity at east δ, i i={,,...,k},j Si ϵ ( à m ) d, where ϵ dm = C 5 fk( + η) ϵ(ϵg + η + h ). Proof. Consider any i {,,..., k} and j S i. Let j = arg min j Si S i L j L j. We wi consider the expression: [ : S i à j ] à m d dm () The first term in the above expression can be upper-bounded by Jensen s inequaity as foows : S i à j = : S i à j = : S i à j n i : S i (Ãj The right-hand side of Equation 3 can be rewritten and bounded as n i : S i ( à j d d ) n i β : S i ( à j d ) n i β : S i ). (3) ( ) à j d d kβϵ. j (4) Combining these resuts together and appying Equation we see that, with probabiity at east δ, : S i à j kβϵ. (5) Now focus on bounding the second term in Expression. Reca that L m L m g n L L F. (6)

13 Fast Spectra Custering via the Nyström Method 3 Simiary, as in previous paragraph, we can write that L m L m = = à m d() d (m) à m d() d (m) A m d() d (m) n i, (7) The ast equaity uses the fact that,m Si A m = and d () = d (m) = d = d m = n i. We can then expand the right-hand side of Equation 7: à m d() d (m) = n i = à m d dm Equation 8 can be ower-bounded as: à m d dm d dm d () d(m) α d dm d () d(m) Ãm n i d() d(m) Combining Equation 6 and 9 we obtain: à m d () d(m) Ãm n i d() d(m) + α Ãm n i d() d(m) + n i +. (8) à m d dm ni ñ i + à m d dm h + (9) à m d dm α ( L L F + h ). () After appying Theorem 6 we obtain that with probabiity at east δ Combining Equation 5 and we get the foowing: à m d dm α ( g n ϵn + η + h ). () [ : S i à j ] à m d dm kβϵ α (ϵg + η + h ) C 5 fk( + η) ϵ(ϵg + η + h ) () where the ast inequaity uses the fact that α ( d D ) and β D d.

14 4 A. Choromanska et a. 5 Experiments To evauate the proposed agorithms empiricay, we consider the four datasets described in Ng et a. (). We used a Gaussian kerne to buid the affinity matrix (κ(s i, s j ) = exp( s i s j /σ )). The parameters σ and r were manuay tuned to obtain the best performance. Figure shows the datasets with pots of the error versus the percent of the coumns samped (/n). We used uniform samping without repacement throughout. Note that both the choice of coumns as we as the initiaization of the k-means custering agorithm sighty affect the performance. Thus, we show two types of resuts: the curves in the second row on Figure obtained by averaging over, runs and the curves underneath showing the most frequenty obtained performance (i.e. the median case). Aso we performed two sets of experiments where r was hed constant as we as where r was tuned for each vaue of. In the first case, we set r = τ (the vaue of τ for each dataset is provided under Figure ) and when τ we set τ =. In the second case, we observed that tuning r for each vaue of (when increases, r shoud decrease) can improve the performance but the improvement is reativey sma and not worth presenting here Error vs. /n 8 Error vs. /n Error vs. /n 5 Error vs. /n % of coumns samped (/n) % of coumns samped (/n) % of coumns samped (/n) % of coumns samped (/n) 5 Error vs. /n Error vs. /n Error vs. /n 6 Error vs. /n % of coumns samped (/n) % of coumns samped (/n) % of coumns samped (/n) % of coumns samped (/n) Fig.. Top row: the datasets with coor-coded custers. Second Row: error curves vs % of coumns samped with the error averaged over, runs. Third row: error curves vs % of coumns samped with the most frequent resut being dispayed. The parameters of interest for each experiment (from eft to right) were: a) n = ; σ = ; τ = 5, b) n = 5; σ = ; τ =, c) n = ; σ = ; τ = 5, d) n = ; σ = ; τ = 5. There was no significant difference in the choice of the distortion-minimizing agorithm we use in the ast step of our spectra custering agorithm, be it Loyd s agorithm, k-means++ and k-means#.

15 Fast Spectra Custering via the Nyström Method 5 Acknowedgments. The authors thank Sanjiv Kumar for hepfu suggestions. References Bekin, M. and Niyogi, P. (7). Convergence of Lapacian eigenmaps. In NIPS 6, pages MIT Press. Drineas, P. and Mahoney, M. W. (5). On the Nyström Method for Approximating a Gram Matrix for Improved Kerne-Based Learning. Journa of Machine Learning Research, 6:5. Fowkes, C., Beongie, S., Chung, F., and Maik, J. (4). Spectra grouping using the nyström method. IEEE Trans. Pattern Ana. Mach. Inte., 6():4 5. Fung, W. S., Hariharan, R., Harvey, N. J., and Panigrahi, D. (). A genera framework for graph sparsification. In STOC. Kannan, R. and Vempaa, S. (9). Spectra agorithms. Foundations and Trends in Theoretica Computer Science, 4(3-4): Kumar, S., Mohri, M., and Tawakar, A. (9). Samping techniques for the nyström method. Journa of Machine Learning Research, 5:34 3. Lashkari, D. and Goand, P. (7). Convex custering with exempar-based modes. In NIPS 7. Li, M., Lian, X.-C., Kwok, J. T., and Lu, B.-L. (). Time and space efficient spectra custering via coumn samping. In 4th IEEE Conference on Computer Vision and Pattern Recognition, CVPR, pages IEEE. Loyd, S. P. (98). Least squares quantization in pcm. IEEE Transactions on Information Theory, 8:9 37. Luxburg, U. (7). A tutoria on spectra custering. Statistics and Computing, 7(4): Ng, A. Y., Jordan, M. I., and Weiss, Y. (). On spectra custering: Anaysis and an agorithm. In NIPS, pages MIT Press. Spieman, D. A. and Teng, S.-H. (). Spectra sparsification of graphs. SIAM Journa on Computing, 4(4):98 5. Wiiams, C. and Seeger, M. (). Using the Nyström method to speed up kerne machines. In NIPS, pages MIT Press. Yan, D., Huang, L., and Jordan, M. I. (9). Fast approximate spectra custering. In ACM SIGKDD, pages ACM.

SUPPLEMENTARY MATERIAL TO INNOVATED SCALABLE EFFICIENT ESTIMATION IN ULTRA-LARGE GAUSSIAN GRAPHICAL MODELS

SUPPLEMENTARY MATERIAL TO INNOVATED SCALABLE EFFICIENT ESTIMATION IN ULTRA-LARGE GAUSSIAN GRAPHICAL MODELS ISEE 1 SUPPLEMENTARY MATERIAL TO INNOVATED SCALABLE EFFICIENT ESTIMATION IN ULTRA-LARGE GAUSSIAN GRAPHICAL MODELS By Yingying Fan and Jinchi Lv University of Southern Caifornia This Suppementary Materia

More information

A Brief Introduction to Markov Chains and Hidden Markov Models

A Brief Introduction to Markov Chains and Hidden Markov Models A Brief Introduction to Markov Chains and Hidden Markov Modes Aen B MacKenzie Notes for December 1, 3, &8, 2015 Discrete-Time Markov Chains You may reca that when we first introduced random processes,

More information

MARKOV CHAINS AND MARKOV DECISION THEORY. Contents

MARKOV CHAINS AND MARKOV DECISION THEORY. Contents MARKOV CHAINS AND MARKOV DECISION THEORY ARINDRIMA DATTA Abstract. In this paper, we begin with a forma introduction to probabiity and expain the concept of random variabes and stochastic processes. After

More information

FRST Multivariate Statistics. Multivariate Discriminant Analysis (MDA)

FRST Multivariate Statistics. Multivariate Discriminant Analysis (MDA) 1 FRST 531 -- Mutivariate Statistics Mutivariate Discriminant Anaysis (MDA) Purpose: 1. To predict which group (Y) an observation beongs to based on the characteristics of p predictor (X) variabes, using

More information

Lecture Note 3: Stationary Iterative Methods

Lecture Note 3: Stationary Iterative Methods MATH 5330: Computationa Methods of Linear Agebra Lecture Note 3: Stationary Iterative Methods Xianyi Zeng Department of Mathematica Sciences, UTEP Stationary Iterative Methods The Gaussian eimination (or

More information

Distributed average consensus: Beyond the realm of linearity

Distributed average consensus: Beyond the realm of linearity Distributed average consensus: Beyond the ream of inearity Usman A. Khan, Soummya Kar, and José M. F. Moura Department of Eectrica and Computer Engineering Carnegie Meon University 5 Forbes Ave, Pittsburgh,

More information

8 APPENDIX. E[m M] = (n S )(1 exp( exp(s min + c M))) (19) E[m M] n exp(s min + c M) (20) 8.1 EMPIRICAL EVALUATION OF SAMPLING

8 APPENDIX. E[m M] = (n S )(1 exp( exp(s min + c M))) (19) E[m M] n exp(s min + c M) (20) 8.1 EMPIRICAL EVALUATION OF SAMPLING 8 APPENDIX 8.1 EMPIRICAL EVALUATION OF SAMPLING We wish to evauate the empirica accuracy of our samping technique on concrete exampes. We do this in two ways. First, we can sort the eements by probabiity

More information

arxiv: v1 [cs.lg] 31 Oct 2017

arxiv: v1 [cs.lg] 31 Oct 2017 ACCELERATED SPARSE SUBSPACE CLUSTERING Abofaz Hashemi and Haris Vikao Department of Eectrica and Computer Engineering, University of Texas at Austin, Austin, TX, USA arxiv:7.26v [cs.lg] 3 Oct 27 ABSTRACT

More information

An explicit Jordan Decomposition of Companion matrices

An explicit Jordan Decomposition of Companion matrices An expicit Jordan Decomposition of Companion matrices Fermín S V Bazán Departamento de Matemática CFM UFSC 88040-900 Forianópois SC E-mai: fermin@mtmufscbr S Gratton CERFACS 42 Av Gaspard Coriois 31057

More information

Explicit overall risk minimization transductive bound

Explicit overall risk minimization transductive bound 1 Expicit overa risk minimization transductive bound Sergio Decherchi, Paoo Gastado, Sandro Ridea, Rodofo Zunino Dept. of Biophysica and Eectronic Engineering (DIBE), Genoa University Via Opera Pia 11a,

More information

Bayesian Learning. You hear a which which could equally be Thanks or Tanks, which would you go with?

Bayesian Learning. You hear a which which could equally be Thanks or Tanks, which would you go with? Bayesian Learning A powerfu and growing approach in machine earning We use it in our own decision making a the time You hear a which which coud equay be Thanks or Tanks, which woud you go with? Combine

More information

CS229 Lecture notes. Andrew Ng

CS229 Lecture notes. Andrew Ng CS229 Lecture notes Andrew Ng Part IX The EM agorithm In the previous set of notes, we taked about the EM agorithm as appied to fitting a mixture of Gaussians. In this set of notes, we give a broader view

More information

BALANCING REGULAR MATRIX PENCILS

BALANCING REGULAR MATRIX PENCILS BALANCING REGULAR MATRIX PENCILS DAMIEN LEMONNIER AND PAUL VAN DOOREN Abstract. In this paper we present a new diagona baancing technique for reguar matrix pencis λb A, which aims at reducing the sensitivity

More information

Problem set 6 The Perron Frobenius theorem.

Problem set 6 The Perron Frobenius theorem. Probem set 6 The Perron Frobenius theorem. Math 22a4 Oct 2 204, Due Oct.28 In a future probem set I want to discuss some criteria which aow us to concude that that the ground state of a sef-adjoint operator

More information

Investigation on spectrum of the adjacency matrix and Laplacian matrix of graph G l

Investigation on spectrum of the adjacency matrix and Laplacian matrix of graph G l Investigation on spectrum of the adjacency matrix and Lapacian matrix of graph G SHUHUA YIN Computer Science and Information Technoogy Coege Zhejiang Wani University Ningbo 3500 PEOPLE S REPUBLIC OF CHINA

More information

MATH 172: MOTIVATION FOR FOURIER SERIES: SEPARATION OF VARIABLES

MATH 172: MOTIVATION FOR FOURIER SERIES: SEPARATION OF VARIABLES MATH 172: MOTIVATION FOR FOURIER SERIES: SEPARATION OF VARIABLES Separation of variabes is a method to sove certain PDEs which have a warped product structure. First, on R n, a inear PDE of order m is

More information

Preconditioned Locally Harmonic Residual Method for Computing Interior Eigenpairs of Certain Classes of Hermitian Matrices

Preconditioned Locally Harmonic Residual Method for Computing Interior Eigenpairs of Certain Classes of Hermitian Matrices MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.mer.com Preconditioned Locay Harmonic Residua Method for Computing Interior Eigenpairs of Certain Casses of Hermitian Matrices Vecharynski, E.; Knyazev,

More information

Appendix of the Paper The Role of No-Arbitrage on Forecasting: Lessons from a Parametric Term Structure Model

Appendix of the Paper The Role of No-Arbitrage on Forecasting: Lessons from a Parametric Term Structure Model Appendix of the Paper The Roe of No-Arbitrage on Forecasting: Lessons from a Parametric Term Structure Mode Caio Ameida cameida@fgv.br José Vicente jose.vaentim@bcb.gov.br June 008 1 Introduction In this

More information

Absolute Value Preconditioning for Symmetric Indefinite Linear Systems

Absolute Value Preconditioning for Symmetric Indefinite Linear Systems MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.mer.com Absoute Vaue Preconditioning for Symmetric Indefinite Linear Systems Vecharynski, E.; Knyazev, A.V. TR2013-016 March 2013 Abstract We introduce

More information

A. Distribution of the test statistic

A. Distribution of the test statistic A. Distribution of the test statistic In the sequentia test, we first compute the test statistic from a mini-batch of size m. If a decision cannot be made with this statistic, we keep increasing the mini-batch

More information

arxiv: v1 [cs.ds] 12 Nov 2018

arxiv: v1 [cs.ds] 12 Nov 2018 Quantum-inspired ow-rank stochastic regression with ogarithmic dependence on the dimension András Giyén 1, Seth Loyd Ewin Tang 3 November 13, 018 arxiv:181104909v1 [csds] 1 Nov 018 Abstract We construct

More information

Stochastic Complement Analysis of Multi-Server Threshold Queues. with Hysteresis. Abstract

Stochastic Complement Analysis of Multi-Server Threshold Queues. with Hysteresis. Abstract Stochastic Compement Anaysis of Muti-Server Threshod Queues with Hysteresis John C.S. Lui The Dept. of Computer Science & Engineering The Chinese University of Hong Kong Leana Goubchik Dept. of Computer

More information

C. Fourier Sine Series Overview

C. Fourier Sine Series Overview 12 PHILIP D. LOEWEN C. Fourier Sine Series Overview Let some constant > be given. The symboic form of the FSS Eigenvaue probem combines an ordinary differentia equation (ODE) on the interva (, ) with a

More information

(This is a sample cover image for this issue. The actual cover is not yet available at this time.)

(This is a sample cover image for this issue. The actual cover is not yet available at this time.) (This is a sampe cover image for this issue The actua cover is not yet avaiabe at this time) This artice appeared in a journa pubished by Esevier The attached copy is furnished to the author for interna

More information

Two view learning: SVM-2K, Theory and Practice

Two view learning: SVM-2K, Theory and Practice Two view earning: SVM-2K, Theory and Practice Jason D.R. Farquhar jdrf99r@ecs.soton.ac.uk Hongying Meng hongying@cs.york.ac.uk David R. Hardoon drh@ecs.soton.ac.uk John Shawe-Tayor jst@ecs.soton.ac.uk

More information

Expectation-Maximization for Estimating Parameters for a Mixture of Poissons

Expectation-Maximization for Estimating Parameters for a Mixture of Poissons Expectation-Maximization for Estimating Parameters for a Mixture of Poissons Brandon Maone Department of Computer Science University of Hesini February 18, 2014 Abstract This document derives, in excrutiating

More information

Math 124B January 31, 2012

Math 124B January 31, 2012 Math 124B January 31, 212 Viktor Grigoryan 7 Inhomogeneous boundary vaue probems Having studied the theory of Fourier series, with which we successfuy soved boundary vaue probems for the homogeneous heat

More information

Statistical Learning Theory: A Primer

Statistical Learning Theory: A Primer Internationa Journa of Computer Vision 38(), 9 3, 2000 c 2000 uwer Academic Pubishers. Manufactured in The Netherands. Statistica Learning Theory: A Primer THEODOROS EVGENIOU, MASSIMILIANO PONTIL AND TOMASO

More information

XSAT of linear CNF formulas

XSAT of linear CNF formulas XSAT of inear CN formuas Bernd R. Schuh Dr. Bernd Schuh, D-50968 Kön, Germany; bernd.schuh@netcoogne.de eywords: compexity, XSAT, exact inear formua, -reguarity, -uniformity, NPcompeteness Abstract. Open

More information

Maximizing Sum Rate and Minimizing MSE on Multiuser Downlink: Optimality, Fast Algorithms and Equivalence via Max-min SIR

Maximizing Sum Rate and Minimizing MSE on Multiuser Downlink: Optimality, Fast Algorithms and Equivalence via Max-min SIR 1 Maximizing Sum Rate and Minimizing MSE on Mutiuser Downink: Optimaity, Fast Agorithms and Equivaence via Max-min SIR Chee Wei Tan 1,2, Mung Chiang 2 and R. Srikant 3 1 Caifornia Institute of Technoogy,

More information

Active Learning & Experimental Design

Active Learning & Experimental Design Active Learning & Experimenta Design Danie Ting Heaviy modified, of course, by Lye Ungar Origina Sides by Barbara Engehardt and Aex Shyr Lye Ungar, University of Pennsyvania Motivation u Data coection

More information

Smoothness equivalence properties of univariate subdivision schemes and their projection analogues

Smoothness equivalence properties of univariate subdivision schemes and their projection analogues Numerische Mathematik manuscript No. (wi be inserted by the editor) Smoothness equivaence properties of univariate subdivision schemes and their projection anaogues Phiipp Grohs TU Graz Institute of Geometry

More information

Source and Relay Matrices Optimization for Multiuser Multi-Hop MIMO Relay Systems

Source and Relay Matrices Optimization for Multiuser Multi-Hop MIMO Relay Systems Source and Reay Matrices Optimization for Mutiuser Muti-Hop MIMO Reay Systems Yue Rong Department of Eectrica and Computer Engineering, Curtin University, Bentey, WA 6102, Austraia Abstract In this paper,

More information

Mat 1501 lecture notes, penultimate installment

Mat 1501 lecture notes, penultimate installment Mat 1501 ecture notes, penutimate instament 1. bounded variation: functions of a singe variabe optiona) I beieve that we wi not actuay use the materia in this section the point is mainy to motivate the

More information

ORTHOGONAL MULTI-WAVELETS FROM MATRIX FACTORIZATION

ORTHOGONAL MULTI-WAVELETS FROM MATRIX FACTORIZATION J. Korean Math. Soc. 46 2009, No. 2, pp. 281 294 ORHOGONAL MLI-WAVELES FROM MARIX FACORIZAION Hongying Xiao Abstract. Accuracy of the scaing function is very crucia in waveet theory, or correspondingy,

More information

Stochastic Variational Inference with Gradient Linearization

Stochastic Variational Inference with Gradient Linearization Stochastic Variationa Inference with Gradient Linearization Suppementa Materia Tobias Pötz * Anne S Wannenwetsch Stefan Roth Department of Computer Science, TU Darmstadt Preface In this suppementa materia,

More information

Sequential Decoding of Polar Codes with Arbitrary Binary Kernel

Sequential Decoding of Polar Codes with Arbitrary Binary Kernel Sequentia Decoding of Poar Codes with Arbitrary Binary Kerne Vera Miosavskaya, Peter Trifonov Saint-Petersburg State Poytechnic University Emai: veram,petert}@dcn.icc.spbstu.ru Abstract The probem of efficient

More information

Cryptanalysis of PKP: A New Approach

Cryptanalysis of PKP: A New Approach Cryptanaysis of PKP: A New Approach Éiane Jaumes and Antoine Joux DCSSI 18, rue du Dr. Zamenhoff F-92131 Issy-es-Mx Cedex France eiane.jaumes@wanadoo.fr Antoine.Joux@ens.fr Abstract. Quite recenty, in

More information

Partial permutation decoding for MacDonald codes

Partial permutation decoding for MacDonald codes Partia permutation decoding for MacDonad codes J.D. Key Department of Mathematics and Appied Mathematics University of the Western Cape 7535 Bevie, South Africa P. Seneviratne Department of Mathematics

More information

Akaike Information Criterion for ANOVA Model with a Simple Order Restriction

Akaike Information Criterion for ANOVA Model with a Simple Order Restriction Akaike Information Criterion for ANOVA Mode with a Simpe Order Restriction Yu Inatsu * Department of Mathematics, Graduate Schoo of Science, Hiroshima University ABSTRACT In this paper, we consider Akaike

More information

Moreau-Yosida Regularization for Grouped Tree Structure Learning

Moreau-Yosida Regularization for Grouped Tree Structure Learning Moreau-Yosida Reguarization for Grouped Tree Structure Learning Jun Liu Computer Science and Engineering Arizona State University J.Liu@asu.edu Jieping Ye Computer Science and Engineering Arizona State

More information

THE REACHABILITY CONES OF ESSENTIALLY NONNEGATIVE MATRICES

THE REACHABILITY CONES OF ESSENTIALLY NONNEGATIVE MATRICES THE REACHABILITY CONES OF ESSENTIALLY NONNEGATIVE MATRICES by Michae Neumann Department of Mathematics, University of Connecticut, Storrs, CT 06269 3009 and Ronad J. Stern Department of Mathematics, Concordia

More information

4 Separation of Variables

4 Separation of Variables 4 Separation of Variabes In this chapter we describe a cassica technique for constructing forma soutions to inear boundary vaue probems. The soution of three cassica (paraboic, hyperboic and eiptic) PDE

More information

First-Order Corrections to Gutzwiller s Trace Formula for Systems with Discrete Symmetries

First-Order Corrections to Gutzwiller s Trace Formula for Systems with Discrete Symmetries c 26 Noninear Phenomena in Compex Systems First-Order Corrections to Gutzwier s Trace Formua for Systems with Discrete Symmetries Hoger Cartarius, Jörg Main, and Günter Wunner Institut für Theoretische

More information

A proposed nonparametric mixture density estimation using B-spline functions

A proposed nonparametric mixture density estimation using B-spline functions A proposed nonparametric mixture density estimation using B-spine functions Atizez Hadrich a,b, Mourad Zribi a, Afif Masmoudi b a Laboratoire d Informatique Signa et Image de a Côte d Opae (LISIC-EA 4491),

More information

LECTURE NOTES 8 THE TRACELESS SYMMETRIC TENSOR EXPANSION AND STANDARD SPHERICAL HARMONICS

LECTURE NOTES 8 THE TRACELESS SYMMETRIC TENSOR EXPANSION AND STANDARD SPHERICAL HARMONICS MASSACHUSETTS INSTITUTE OF TECHNOLOGY Physics Department Physics 8.07: Eectromagnetism II October, 202 Prof. Aan Guth LECTURE NOTES 8 THE TRACELESS SYMMETRIC TENSOR EXPANSION AND STANDARD SPHERICAL HARMONICS

More information

ASummaryofGaussianProcesses Coryn A.L. Bailer-Jones

ASummaryofGaussianProcesses Coryn A.L. Bailer-Jones ASummaryofGaussianProcesses Coryn A.L. Baier-Jones Cavendish Laboratory University of Cambridge caj@mrao.cam.ac.uk Introduction A genera prediction probem can be posed as foows. We consider that the variabe

More information

Separation of Variables and a Spherical Shell with Surface Charge

Separation of Variables and a Spherical Shell with Surface Charge Separation of Variabes and a Spherica She with Surface Charge In cass we worked out the eectrostatic potentia due to a spherica she of radius R with a surface charge density σθ = σ cos θ. This cacuation

More information

Soft Clustering on Graphs

Soft Clustering on Graphs Soft Custering on Graphs Kai Yu 1, Shipeng Yu 2, Voker Tresp 1 1 Siemens AG, Corporate Technoogy 2 Institute for Computer Science, University of Munich kai.yu@siemens.com, voker.tresp@siemens.com spyu@dbs.informatik.uni-muenchen.de

More information

Week 6 Lectures, Math 6451, Tanveer

Week 6 Lectures, Math 6451, Tanveer Fourier Series Week 6 Lectures, Math 645, Tanveer In the context of separation of variabe to find soutions of PDEs, we encountered or and in other cases f(x = f(x = a 0 + f(x = a 0 + b n sin nπx { a n

More information

Efficiently Generating Random Bits from Finite State Markov Chains

Efficiently Generating Random Bits from Finite State Markov Chains 1 Efficienty Generating Random Bits from Finite State Markov Chains Hongchao Zhou and Jehoshua Bruck, Feow, IEEE Abstract The probem of random number generation from an uncorreated random source (of unknown

More information

arxiv: v1 [cs.db] 1 Aug 2012

arxiv: v1 [cs.db] 1 Aug 2012 Functiona Mechanism: Regression Anaysis under Differentia Privacy arxiv:208.029v [cs.db] Aug 202 Jun Zhang Zhenjie Zhang 2 Xiaokui Xiao Yin Yang 2 Marianne Winsett 2,3 ABSTRACT Schoo of Computer Engineering

More information

Restricted weak type on maximal linear and multilinear integral maps.

Restricted weak type on maximal linear and multilinear integral maps. Restricted weak type on maxima inear and mutiinear integra maps. Oscar Basco Abstract It is shown that mutiinear operators of the form T (f 1,..., f k )(x) = R K(x, y n 1,..., y k )f 1 (y 1 )...f k (y

More information

Statistics for Applications. Chapter 7: Regression 1/43

Statistics for Applications. Chapter 7: Regression 1/43 Statistics for Appications Chapter 7: Regression 1/43 Heuristics of the inear regression (1) Consider a coud of i.i.d. random points (X i,y i ),i =1,...,n : 2/43 Heuristics of the inear regression (2)

More information

Multilayer Kerceptron

Multilayer Kerceptron Mutiayer Kerceptron Zotán Szabó, András Lőrincz Department of Information Systems, Facuty of Informatics Eötvös Loránd University Pázmány Péter sétány 1/C H-1117, Budapest, Hungary e-mai: szzoi@csetehu,

More information

FFTs in Graphics and Vision. Spherical Convolution and Axial Symmetry Detection

FFTs in Graphics and Vision. Spherical Convolution and Axial Symmetry Detection FFTs in Graphics and Vision Spherica Convoution and Axia Symmetry Detection Outine Math Review Symmetry Genera Convoution Spherica Convoution Axia Symmetry Detection Math Review Symmetry: Given a unitary

More information

Mode in Output Participation Factors for Linear Systems

Mode in Output Participation Factors for Linear Systems 2010 American ontro onference Marriott Waterfront, Batimore, MD, USA June 30-Juy 02, 2010 WeB05.5 Mode in Output Participation Factors for Linear Systems Li Sheng, yad H. Abed, Munther A. Hassouneh, Huizhong

More information

LECTURE NOTES 9 TRACELESS SYMMETRIC TENSOR APPROACH TO LEGENDRE POLYNOMIALS AND SPHERICAL HARMONICS

LECTURE NOTES 9 TRACELESS SYMMETRIC TENSOR APPROACH TO LEGENDRE POLYNOMIALS AND SPHERICAL HARMONICS MASSACHUSETTS INSTITUTE OF TECHNOLOGY Physics Department Physics 8.07: Eectromagnetism II October 7, 202 Prof. Aan Guth LECTURE NOTES 9 TRACELESS SYMMETRIC TENSOR APPROACH TO LEGENDRE POLYNOMIALS AND SPHERICAL

More information

arxiv: v1 [math.co] 17 Dec 2018

arxiv: v1 [math.co] 17 Dec 2018 On the Extrema Maximum Agreement Subtree Probem arxiv:1812.06951v1 [math.o] 17 Dec 2018 Aexey Markin Department of omputer Science, Iowa State University, USA amarkin@iastate.edu Abstract Given two phyogenetic

More information

STA 216 Project: Spline Approach to Discrete Survival Analysis

STA 216 Project: Spline Approach to Discrete Survival Analysis : Spine Approach to Discrete Surviva Anaysis November 4, 005 1 Introduction Athough continuous surviva anaysis differs much from the discrete surviva anaysis, there is certain ink between the two modeing

More information

II. PROBLEM. A. Description. For the space of audio signals

II. PROBLEM. A. Description. For the space of audio signals CS229 - Fina Report Speech Recording based Language Recognition (Natura Language) Leopod Cambier - cambier; Matan Leibovich - matane; Cindy Orozco Bohorquez - orozcocc ABSTRACT We construct a rea time

More information

Uniprocessor Feasibility of Sporadic Tasks with Constrained Deadlines is Strongly conp-complete

Uniprocessor Feasibility of Sporadic Tasks with Constrained Deadlines is Strongly conp-complete Uniprocessor Feasibiity of Sporadic Tasks with Constrained Deadines is Strongy conp-compete Pontus Ekberg and Wang Yi Uppsaa University, Sweden Emai: {pontus.ekberg yi}@it.uu.se Abstract Deciding the feasibiity

More information

New Efficiency Results for Makespan Cost Sharing

New Efficiency Results for Makespan Cost Sharing New Efficiency Resuts for Makespan Cost Sharing Yvonne Beischwitz a, Forian Schoppmann a, a University of Paderborn, Department of Computer Science Fürstenaee, 3302 Paderborn, Germany Abstract In the context

More information

Optimality of Inference in Hierarchical Coding for Distributed Object-Based Representations

Optimality of Inference in Hierarchical Coding for Distributed Object-Based Representations Optimaity of Inference in Hierarchica Coding for Distributed Object-Based Representations Simon Brodeur, Jean Rouat NECOTIS, Département génie éectrique et génie informatique, Université de Sherbrooke,

More information

6.434J/16.391J Statistics for Engineers and Scientists May 4 MIT, Spring 2006 Handout #17. Solution 7

6.434J/16.391J Statistics for Engineers and Scientists May 4 MIT, Spring 2006 Handout #17. Solution 7 6.434J/16.391J Statistics for Engineers and Scientists May 4 MIT, Spring 2006 Handout #17 Soution 7 Probem 1: Generating Random Variabes Each part of this probem requires impementation in MATLAB. For the

More information

B. Brown, M. Griebel, F.Y. Kuo and I.H. Sloan

B. Brown, M. Griebel, F.Y. Kuo and I.H. Sloan Wegeerstraße 6 53115 Bonn Germany phone +49 8 73-347 fax +49 8 73-757 www.ins.uni-bonn.de B. Brown, M. Griebe, F.Y. Kuo and I.H. Soan On the expected uniform error of geometric Brownian motion approximated

More information

arxiv: v1 [math.fa] 23 Aug 2018

arxiv: v1 [math.fa] 23 Aug 2018 An Exact Upper Bound on the L p Lebesgue Constant and The -Rényi Entropy Power Inequaity for Integer Vaued Random Variabes arxiv:808.0773v [math.fa] 3 Aug 08 Peng Xu, Mokshay Madiman, James Mebourne Abstract

More information

Higher dimensional PDEs and multidimensional eigenvalue problems

Higher dimensional PDEs and multidimensional eigenvalue problems Higher dimensiona PEs and mutidimensiona eigenvaue probems 1 Probems with three independent variabes Consider the prototypica equations u t = u (iffusion) u tt = u (W ave) u zz = u (Lapace) where u = u

More information

Alberto Maydeu Olivares Instituto de Empresa Marketing Dept. C/Maria de Molina Madrid Spain

Alberto Maydeu Olivares Instituto de Empresa Marketing Dept. C/Maria de Molina Madrid Spain CORRECTIONS TO CLASSICAL PROCEDURES FOR ESTIMATING THURSTONE S CASE V MODEL FOR RANKING DATA Aberto Maydeu Oivares Instituto de Empresa Marketing Dept. C/Maria de Moina -5 28006 Madrid Spain Aberto.Maydeu@ie.edu

More information

A SIMPLIFIED DESIGN OF MULTIDIMENSIONAL TRANSFER FUNCTION MODELS

A SIMPLIFIED DESIGN OF MULTIDIMENSIONAL TRANSFER FUNCTION MODELS A SIPLIFIED DESIGN OF ULTIDIENSIONAL TRANSFER FUNCTION ODELS Stefan Petrausch, Rudof Rabenstein utimedia Communications and Signa Procesg, University of Erangen-Nuremberg, Cauerstr. 7, 958 Erangen, GERANY

More information

On Non-Optimally Expanding Sets in Grassmann Graphs

On Non-Optimally Expanding Sets in Grassmann Graphs ectronic Cooquium on Computationa Compexity, Report No. 94 (07) On Non-Optimay xpanding Sets in Grassmann Graphs Irit Dinur Subhash Khot Guy Kinder Dor Minzer Mui Safra Abstract The paper investigates

More information

DIGITAL FILTER DESIGN OF IIR FILTERS USING REAL VALUED GENETIC ALGORITHM

DIGITAL FILTER DESIGN OF IIR FILTERS USING REAL VALUED GENETIC ALGORITHM DIGITAL FILTER DESIGN OF IIR FILTERS USING REAL VALUED GENETIC ALGORITHM MIKAEL NILSSON, MATTIAS DAHL AND INGVAR CLAESSON Bekinge Institute of Technoogy Department of Teecommunications and Signa Processing

More information

Math 124B January 17, 2012

Math 124B January 17, 2012 Math 124B January 17, 212 Viktor Grigoryan 3 Fu Fourier series We saw in previous ectures how the Dirichet and Neumann boundary conditions ead to respectivey sine and cosine Fourier series of the initia

More information

6 Wave Equation on an Interval: Separation of Variables

6 Wave Equation on an Interval: Separation of Variables 6 Wave Equation on an Interva: Separation of Variabes 6.1 Dirichet Boundary Conditions Ref: Strauss, Chapter 4 We now use the separation of variabes technique to study the wave equation on a finite interva.

More information

Limits on Support Recovery with Probabilistic Models: An Information-Theoretic Framework

Limits on Support Recovery with Probabilistic Models: An Information-Theoretic Framework Limits on Support Recovery with Probabiistic Modes: An Information-Theoretic Framewor Jonathan Scarett and Voan Cevher arxiv:5.744v3 cs.it 3 Aug 6 Abstract The support recovery probem consists of determining

More information

Substructuring Preconditioners for the Bidomain Extracellular Potential Problem

Substructuring Preconditioners for the Bidomain Extracellular Potential Problem Substructuring Preconditioners for the Bidomain Extraceuar Potentia Probem Mico Pennacchio 1 and Vaeria Simoncini 2,1 1 IMATI - CNR, via Ferrata, 1, 27100 Pavia, Itay mico@imaticnrit 2 Dipartimento di

More information

A Simple and Efficient Algorithm of 3-D Single-Source Localization with Uniform Cross Array Bing Xue 1 2 a) * Guangyou Fang 1 2 b and Yicai Ji 1 2 c)

A Simple and Efficient Algorithm of 3-D Single-Source Localization with Uniform Cross Array Bing Xue 1 2 a) * Guangyou Fang 1 2 b and Yicai Ji 1 2 c) A Simpe Efficient Agorithm of 3-D Singe-Source Locaization with Uniform Cross Array Bing Xue a * Guangyou Fang b Yicai Ji c Key Laboratory of Eectromagnetic Radiation Sensing Technoogy, Institute of Eectronics,

More information

Fitting Algorithms for MMPP ATM Traffic Models

Fitting Algorithms for MMPP ATM Traffic Models Fitting Agorithms for PP AT Traffic odes A. Nogueira, P. Savador, R. Vaadas University of Aveiro / Institute of Teecommunications, 38-93 Aveiro, Portuga; e-mai: (nogueira, savador, rv)@av.it.pt ABSTRACT

More information

Appendix for Stochastic Gradient Monomial Gamma Sampler

Appendix for Stochastic Gradient Monomial Gamma Sampler 3 4 5 6 7 8 9 3 4 5 6 7 8 9 3 4 5 6 7 8 9 3 3 3 33 34 35 36 37 38 39 4 4 4 43 44 45 46 47 48 49 5 5 5 53 54 Appendix for Stochastic Gradient Monomia Gamma Samper A The Main Theorem We provide the foowing

More information

2M2. Fourier Series Prof Bill Lionheart

2M2. Fourier Series Prof Bill Lionheart M. Fourier Series Prof Bi Lionheart 1. The Fourier series of the periodic function f(x) with period has the form f(x) = a 0 + ( a n cos πnx + b n sin πnx ). Here the rea numbers a n, b n are caed the Fourier

More information

On the Goal Value of a Boolean Function

On the Goal Value of a Boolean Function On the Goa Vaue of a Booean Function Eric Bach Dept. of CS University of Wisconsin 1210 W. Dayton St. Madison, WI 53706 Lisa Heerstein Dept of CSE NYU Schoo of Engineering 2 Metrotech Center, 10th Foor

More information

A unified framework for Regularization Networks and Support Vector Machines. Theodoros Evgeniou, Massimiliano Pontil, Tomaso Poggio

A unified framework for Regularization Networks and Support Vector Machines. Theodoros Evgeniou, Massimiliano Pontil, Tomaso Poggio MASSACHUSETTS INSTITUTE OF TECHNOLOGY ARTIFICIAL INTELLIGENCE LABORATORY and CENTER FOR BIOLOGICAL AND COMPUTATIONAL LEARNING DEPARTMENT OF BRAIN AND COGNITIVE SCIENCES A.I. Memo No. 1654 March23, 1999

More information

17 Lecture 17: Recombination and Dark Matter Production

17 Lecture 17: Recombination and Dark Matter Production PYS 652: Astrophysics 88 17 Lecture 17: Recombination and Dark Matter Production New ideas pass through three periods: It can t be done. It probaby can be done, but it s not worth doing. I knew it was

More information

8 Digifl'.11 Cth:uits and devices

8 Digifl'.11 Cth:uits and devices 8 Digif'. Cth:uits and devices 8. Introduction In anaog eectronics, votage is a continuous variabe. This is usefu because most physica quantities we encounter are continuous: sound eves, ight intensity,

More information

A Fundamental Storage-Communication Tradeoff in Distributed Computing with Straggling Nodes

A Fundamental Storage-Communication Tradeoff in Distributed Computing with Straggling Nodes A Fundamenta Storage-Communication Tradeoff in Distributed Computing with Stragging odes ifa Yan, Michèe Wigger LTCI, Téécom ParisTech 75013 Paris, France Emai: {qifa.yan, michee.wigger} @teecom-paristech.fr

More information

An implicit Jacobi-like method for computing generalized hyperbolic SVD

An implicit Jacobi-like method for computing generalized hyperbolic SVD Linear Agebra and its Appications 358 (2003) 293 307 wwweseviercom/ocate/aa An impicit Jacobi-ike method for computing generaized hyperboic SVD Adam W Bojanczyk Schoo of Eectrica and Computer Engineering

More information

Haar Decomposition and Reconstruction Algorithms

Haar Decomposition and Reconstruction Algorithms Jim Lambers MAT 773 Fa Semester 018-19 Lecture 15 and 16 Notes These notes correspond to Sections 4.3 and 4.4 in the text. Haar Decomposition and Reconstruction Agorithms Decomposition Suppose we approximate

More information

Asynchronous Control for Coupled Markov Decision Systems

Asynchronous Control for Coupled Markov Decision Systems INFORMATION THEORY WORKSHOP (ITW) 22 Asynchronous Contro for Couped Marov Decision Systems Michae J. Neey University of Southern Caifornia Abstract This paper considers optima contro for a coection of

More information

(f) is called a nearly holomorphic modular form of weight k + 2r as in [5].

(f) is called a nearly holomorphic modular form of weight k + 2r as in [5]. PRODUCTS OF NEARLY HOLOMORPHIC EIGENFORMS JEFFREY BEYERL, KEVIN JAMES, CATHERINE TRENTACOSTE, AND HUI XUE Abstract. We prove that the product of two neary hoomorphic Hece eigenforms is again a Hece eigenform

More information

Emmanuel Abbe Colin Sandon

Emmanuel Abbe Colin Sandon Detection in the stochastic bock mode with mutipe custers: proof of the achievabiity conjectures, acycic BP, and the information-computation gap Emmanue Abbe Coin Sandon Abstract In a paper that initiated

More information

Approximated MLC shape matrix decomposition with interleaf collision constraint

Approximated MLC shape matrix decomposition with interleaf collision constraint Approximated MLC shape matrix decomposition with intereaf coision constraint Thomas Kainowski Antje Kiese Abstract Shape matrix decomposition is a subprobem in radiation therapy panning. A given fuence

More information

Fitting affine and orthogonal transformations between two sets of points

Fitting affine and orthogonal transformations between two sets of points Mathematica Communications 9(2004), 27-34 27 Fitting affine and orthogona transformations between two sets of points Hemuth Späth Abstract. Let two point sets P and Q be given in R n. We determine a transation

More information

Scalable Spectrum Allocation for Large Networks Based on Sparse Optimization

Scalable Spectrum Allocation for Large Networks Based on Sparse Optimization Scaabe Spectrum ocation for Large Networks ased on Sparse Optimization innan Zhuang Modem R&D Lab Samsung Semiconductor, Inc. San Diego, C Dongning Guo, Ermin Wei, and Michae L. Honig Department of Eectrica

More information

MAT 167: Advanced Linear Algebra

MAT 167: Advanced Linear Algebra < Proem 1 (15 pts) MAT 167: Advanced Linear Agera Fina Exam Soutions (a) (5 pts) State the definition of a unitary matrix and expain the difference etween an orthogona matrix and an unitary matrix. Soution:

More information

Introduction. Figure 1 W8LC Line Array, box and horn element. Highlighted section modelled.

Introduction. Figure 1 W8LC Line Array, box and horn element. Highlighted section modelled. imuation of the acoustic fied produced by cavities using the Boundary Eement Rayeigh Integra Method () and its appication to a horn oudspeaer. tephen Kirup East Lancashire Institute, Due treet, Bacburn,

More information

arxiv: v1 [math.pr] 30 Nov 2017

arxiv: v1 [math.pr] 30 Nov 2017 REGULARIZATIO OF O-ORMAL MATRICES BY GAUSSIA OISE - THE BADED TOEPLITZ AD TWISTED TOEPLITZ CASES AIRBA BASAK, ELLIOT PAQUETTE, AD OFER ZEITOUI arxiv:7242v [mathpr] 3 ov 27 Abstract We consider the spectrum

More information

Testing for the Existence of Clusters

Testing for the Existence of Clusters Testing for the Existence of Custers Caudio Fuentes and George Casea University of Forida November 13, 2008 Abstract The detection and determination of custers has been of specia interest, among researchers

More information

FOURIER SERIES ON ANY INTERVAL

FOURIER SERIES ON ANY INTERVAL FOURIER SERIES ON ANY INTERVAL Overview We have spent considerabe time earning how to compute Fourier series for functions that have a period of 2p on the interva (-p,p). We have aso seen how Fourier series

More information

$, (2.1) n="# #. (2.2)

$, (2.1) n=# #. (2.2) Chapter. Eectrostatic II Notes: Most of the materia presented in this chapter is taken from Jackson, Chap.,, and 4, and Di Bartoo, Chap... Mathematica Considerations.. The Fourier series and the Fourier

More information

Appendix for Stochastic Gradient Monomial Gamma Sampler

Appendix for Stochastic Gradient Monomial Gamma Sampler Appendix for Stochastic Gradient Monomia Gamma Samper A The Main Theorem We provide the foowing theorem to characterize the stationary distribution of the stochastic process with SDEs in (3) Theorem 3

More information