Multi-View Clustering via Canonical Correlation Analysis

Size: px
Start display at page:

Download "Multi-View Clustering via Canonical Correlation Analysis"

Transcription

1 Keywors: multi-view learning, clustering, canonical correlation analysis Abstract Clustering ata in high-imensions is believe to be a har problem in general. A number of efficient clustering algorithms evelope in recent years aress this problem by projecting the ata into a lowerimensional subspace, e.g. via Principal Components Analysis PCA) or ranom projections, before clustering. Here, we consier constructing such projections using multiple views of the ata, via Canonical Correlation Analysis CCA). Uner the assumption that conitione on the cluster label the views are uncorrelate, we show that the separation conitions require for the algorithm to be successful are rather mil significantly weaker than prior results in the literature). We provie results for mixture of Gaussians an mixtures of log concave istributions. We also provie empirical support from auio-visual speaker clustering where we esire the clusters to correspon to speaker ID) an from hierarchical Wikipeia ocument clustering where one view is the wors in the ocument an other is the link structure). 1. Introuction The multi-view approach to learning is one in which we have views of the ata sometimes in a rather abstract sense) an the goal is to use the relationship between these views to alleviate the ifficulty of a learning problem of interest [BM98, KF07, AZ07]. In this work, we explore how having two views of the ata makes the clustering problem significantly more tractable. Much recent work has been one on unerstaning uner what conitions we can learn a mixture moel. Preliminary work. Uner review by the International Conference on Machine Learning ICML). Do not istribute. The basic problem is as follows: we are given inepenent samples from a mixture of k istributions, an our task is to either: 1) infer properties of the unerlying mixture moel e.g. the mixing weights, means, etc) or 2) classify a ranom sample accoring to which istribution it was generate from. Uner no restrictions on the unerlying istribution, this problem is consiere to be har. However, in many applications, we are only intereste in clustering the ata when the component istributions are well separate. In fact, the focus of recent clustering algorithms [Das99, VW02, AM05, BV08] is on efficiently learning with as little separation as possible. Typically, these separation conitions are such that when given a ranom sample form the mixture moel, the Bayes optimal classifier is able to reliably, with high probability, recover which cluster generate that point. This work assumes a rather natural multi-view assumption: the assumption is that the views are conitionally) uncorrelate, if we conition on which mixture istribution generate the views. There are many natural applications for which this assumption applies. For example, we can consier multi-moal views, with one view being a vieo stream an the other an auio stream of a speaker here conitione on the speaker ientity an maybe the phoneme both of which coul label the generating cluster), the views may be uncorrelate. A secon example is the wors an link structure in a ocument from a corpus such as Wikipeia here, conitione on the category of each ocument, the wors in it an its link structure may be uncorrelate. In this paper, we provie experiments for both settings. Uner this multi-view assumption, we provie a simple an efficient subspace learning metho, base on Canonical Correlation Analysis CCA). This algorithm is affine invariant an is able to learn with some of the weakest separation conitions to ate. The intuitive reason for this is that uner our multi-view assumption, we are able to approximately) fin the low-imensional subspace spanne by the means of the component istributions. This subspace is important,

2 because, when projecte onto this subspace, the means of the istributions are well-separate, yet the typical istance between points from the same istribution is smaller than in the original space. The number of samples we require to cluster correctly scales as O), where is the ambient imension. Finally, we show through experiments that CCA-base algorithms consistently provie better performance than PCA-base clustering methos when applie to atasets in two ifferent omains auio-visual speaker clustering, an hierarchical Wikipeia ocument clustering by category. Our work shows how the multi-view framework can provie substantial improvements to the clustering problem, aing to the growing boy of results which show how the multi-view framework can alleviate the ifficulty of learning problems. Relate Work. Most provably efficient clustering algorithms first project the ata own to some low imensional space an then cluster the ata in this lower imensional space typically, an algorithm such as single linkage suffices here). Typically, these algorithms also work uner a separation requirement, which is measure by the minimum istance between the means of any two mixture components. One of the first provably efficient algorithms for learning mixture moels is ue to [Das99], who learns a mixture of spherical Gaussians by ranomly projecting the mixture onto a low-imensional subspace. [VW02] provie an algorithm with an improve separation requirement that learns a mixture of k spherical Gaussians, by projecting the mixture own to the k-imensional subspace of highest variance. [KSV05, AM05] exten this result to mixtures of general Gaussians; however, they require a separation proportional to the maximum irectional stanar eviation of any mixture component. [CR08] use a canonical-correlations base algorithm to learn mixtures of axis-aligne Gaussians with a separation proportional to σ, the maximum irectional stanar eviation in the subspace containing the centers of the istributions. Their algorithm requires the coorinate-inepenence property, an an aitional spreaing conition. None of these algorithms are affine invariant. Finally, [BV08] provies an affine-invariant algorithm for learning mixtures of general Gaussians, so long as the mixture has a suitably low Fisher coefficient when in isotropic position. However, their separation involves a rather large polynomial epenence on 1 w min. The two results most closely relate to ours are the work of [VW02] an [CR08]. [VW02] shows that it is sufficient to fin the subspace spanne by the means of the istributions in the mixture for effective clustering. Like our algorithm, [CR08] use a projection onto the top k 1 singular value ecomposition subspace of the canonical correlations matrix. They also require a spreaing conition, which is relate to our requirement on the rank. We borrow techniques from both these papers. [?] propose a similar algorithm for multi-view clustering, in which, ata is projecte onto the top few irections obtaine by a kernel-cca of the multi-view ata. They show empirically, that for clustering images using associate text, where the two views are an image, an text associate with it, an the target clustering is a human-efine category), CCA-base methos outperform PCA-base clustering algorithms. In this paper, we stuy the problem of multi-view clustering. We have ata on a fixe set of objects from two sources, which we call the two views, an our goal is to use this fact to cluster more effectively than with ata from a single source. Following prior theoretical work, our goal is to show that our algorithm recovers the correct clustering when the input obeys certain conitons. This Work. Our input is ata on a fixe set of objects from two views, where View j is generate by a mixture of k Gaussians D j 1,..., Dj k ), for j = 1, 2. To generate a sample, a source i is picke with probability w i, an x 1) an x 2) in Views 1 an 2 are rawn from istributions Di 1 an D2 i. We impose two requirements on these mixtures. First, we require that conitione on the source istribution in the mixture, the two views are uncorrelate. Notice that this is a weaker restriction than the conition that given source i, the samples from Di 1 an D2 i are rawn inepenently. Moreover, this conition allows the istributions in the mixture within each view to be completely general, so long as they are uncorrelate across views. Although we o not show it theoretically, we suspect that our algorithm is robust to small eviations from this assumption. Secon, we require the rank of the CCA matrix across the views to be at least k 1, when each view is in isotropic position, an the k 1-th singular value of this matrix to be at least λ min. This conition ensures that there is sufficient correlation between the views. If the first two conitions hol, then, we can recover the subspace containing the means of the istributions in both views. In aition, for mixtures of Gaussians, if in at least

3 one view, say View 1, we have that for every pair of istributions i an j in the mixture, µ 1 i µ 1 j > Cσ k 1/4 logn/δ) for some constant C, where µ 1 i is the mean of the i-th component istribution in view one an σ is the maximum irectional stanar eviation in the subspace containing the means of the istributions in view 1, then our algorithm can also etermine which istribution each sample came from. Moreover, the number of samples require by us to learn this mixture grows almost) linearly with. This separation conition is consierably weaker than previous results in that σ only epens on the irectional variance in the subspace spanne by the means, which can be consierably lower than the maximum irectional variance over all irections. The only other algorithm which provies affine-invariant guarantess is ue to [BV08] while this result oes not explicitly state results in terms of separation between the means it uses a Fisher coefficient concept), the implie separation is rather large an grows with ecreasing w min, the minimum mixing weight. To get our improve sample complexity bouns, we use a result ue to [RV07], which may be of inepenent interest. We stress that our improve results are really ue the multi-view conition. Ha we simply combine the ata from both views, an applie previous algorithms on the combine ata, we coul not have obtaine our guarantees. We also emphasize that for our algorithm to cluster successfully, it is sufficient for the istributions in the mixture to obey the separation conition in one view, so long as the multiview conition an rank conitions are obeye. Finally, we stuy through experiments, the performance of CCA-base algorithms on ata-sets from two ifferent omains. First, we experiment with auiovisual speaker clustering, in which the two views are auio an face images of a speaker, an the target cluster variable is the speaker. Our experiments show that CCA-base algorithms perform better than PCAbase algorithms on auio ata, just as well on image ata, an are more robust to occlusions an translations of the images. For our secon experiment, we cluster ocuments in Wikipeia. The two views are the wors, an the link structure of a ocument, an the target cluster is the category. Our experiments show that a CCA-base hierarchical clustering algorithm provies much higher performance than PCAbase hierarchical clustering for this ata. 2. The Setting We assume that our ata is generate by a mixture of k istributions. In particular, we assume we obtain samples x = x 1), x 2) ), where x 1) an x 2) are the two views of the ata, which live in the vector spaces V 1 of imension 1 an V 2 of imension 2, respectively. We let = Let µ j i, for i = 1,..., k an j = 1, 2 be the center of istribution i in view j, an let w i be the mixing weight for istribution i. For simplicity, we assume that ata have mean 0. We enote the covariance matrix of the ata as: Σ = E[xx ], Σ 11 = E[x 1) x 1) ) ] Σ 22 = E[x 2) x 2) ) ], Σ 12 = E[x 1) x 2) ) ] Hence, we have: [ ] Σ11 Σ Σ = 21. 1) Σ 12 Σ 22 The multi-view assumption we work with is as follows: Assumption 1 Multi-View Conition) We assume that conitione on the source istribution l in the mixture where l = i is picke with probability w i ), the two views are uncorrelate. More precisely, we assume that for all i [k], E[x 1) x 2) ) l = i] = E[x 1) l = i]e[x 2) ) l = i] This assumption implies that: Σ 12 = i To see this, observe that E[x 1) x 2) ) ] = i = i = i w i µ 1 i µ 2 i ) T. E Di [x 1) x 2) ) ] Pr[D i ] w i E Di [x 1) ] E Di [x 2) ) ] w i µ 1 i µ 2 i ) T 2) As the istributions are in isotropic position, we observe that i w iµ 1 i = i w iµ 2 i = 0. Therefore, the above equation shows that the rank of Σ 12 is at most k 1. We now assume that it has rank precisely k 1. Assumption 2 Non-Degeneracy Conition) We assume that Σ 12 has rank k 1 an that the minimal non-zero singular value of Σ 12 is λ min > 0 where we are working in a coorinate system where Σ 11 an Σ 22 are ientity matrices).

4 For clarity of exposition, we also work in a isotropic coorinate system, in each view. Specifically, the expecte covariance matrix of the ata, in each view, is the ientity matrix, i.e. Σ 11 = I 1, Σ 22 = I 2. As our analysis shows, our algorithm is robust to errors, so we assume that ata is whitene as a preprocessing step. One way to view the Non-Degeneracy Assumption is in terms of correlation coefficients. Recall that for two irections u V 1 an v V 2, the correlation coefficient is efine as: ρu, v) = E[u x 1) )v x 2) )] E[u x 1) ) 2 ]E[v x 2) ) 2 ]. An alternative efinition of λ min is just the minimal non-zero, correlation coefficient i.e. λ min = min ρu, v). u,v:ρu,v) 0 Note 1 λ min > 0.We use Σ 11 an Σ 22 to enote the sample covariance matrix in views 1 an 2 respectively. We use Σ 12 to enote the sample covariance matrix combine across views 1 an 2. We assume these are obtaine through empirical averages from i.i.. samples from the unerlying istribution. 3. The Clustering Algorithm The following lemma provies the intuition for our algorithm. Lemma 1 Uner Assumption 2, if U, D, V is the thin SVD of Σ 12 where the thin SVD removes all zero entries from the iagonal), then the subspace spanne by the means in view 1 is precisely the column span of U an we have the analogous statement for view 2). The lemma is a consequence of Equation 2 an the rank assumption. Since samples from a mixture are well-separate in the space containing the means of the istributions, the lemma suggests the following strategy : use CCA to approximately) project the ata own to the subspace spanne by the means to get an easier clustering problem, an then apply stanar clustering algorithms in this space. Our clustering algorithm, base on the above iea, is state below. We can show that this algorithm clusters correctly with high probability, when the ata in at least one of the views obeys a separation conition, in aition to our assumptions. The input to the algorithm is a set of samples S, an a number k, an the output is a clustering of these samples into k clusters. For this algorithm, we assume that the ata obeys the separation conition in View 1; an analogous algorithm can be applie when the ata obeys the separation conition in View 2 as well. Algorithm Ranomly partition S into two subsets of equal size A an B. 2. Let Σ 12 A) Σ 12 B) respectively) enote the empirical covariance matrix between views 1 an 2, compute from the sample set A B respectively). Compute the top k 1 left singular vectors of Σ 12 A) Σ 12 B) respectively), an project the samples in B A respectively) on the subspace spanne by these vectors. 3. Apply single linkage clustering [?] for mixtures of log-concave istributions), or the algorithm in Section 3.5 of [AK05] for mixtures of Gaussians) on the projecte examples in View 1. We note that in Step 3, we apply either single-linkage or the algorithm of [AK05]; this allows us to show theoretically that given the istributions in the mixture are of a certain type, an given the right separation conitions, the clusters can be recovere correctly. In practice, however, these algorithms o not perform as well ue to lack of robustness, an one woul use an algorithm such as k-means or EM to cluster in this low-imensional subspace. In particular, a variant of the EM algorithm has been shown [DS00] to cluster correctly mixtures of Gaussians, uner certain conitions. Moreover, in Step 1, we ivie the ata-set into two halves to ensure inepenence between Steps 2 an 3 for our analysis; in practice however, these steps can be execute on the same sample set The Main Result Our main theorem can be state as follows. Theorem 1 Gaussians) Suppose the source istribution is a mixture of Gaussians, an suppose Assumptions 1 an 2 hol. Let σ be the maximum irectional stanar eviation of any istribution in the subspace spanne by {µ 1 i }k i=1. If, for each pair i an j an for a fixe constant C, µ 1 i µ 1 j Cσ k 1/4 log kn δ ) then, with probability 1 δ, Algorithm 1 correctly clas-

5 sifies the examples if the number of examples use is c σ ) 2 λ 2 log 2 min w2 min σ ) log 2 1/δ) λ min w min for some constant c. Here we assume that a separation conition hols in View 1, but a similar theorem also applies to View 2. An analogous theorem can also be shown for mixtures of log-concave istributions. Theorem 2 Log-concave Distributions) Suppose the source istribution is a mixture of log-concave istributions, an suppose Assumptions 1 an 2 hol. Let σ be the maximum irectional stanar eviation of any istribution in the subspace spanne by {µ 1 i }k i=1. If, for each pair i an j an for a fixe constant C, µ 1 i µ 1 j Cσ k log kn δ ) then, with probability 1 δ, Algorithm 1 correctly classifies the examples if the number of examples use is c σ ) 2 λ 2 log 3 min w2 min σ ) log 2 1/δ) λ min w min for some constant c. 4. Analyzing Our Algorithm In this section, we prove our main theorems. First, we efine some notation. Notation. In the sequel, we assume that we are given samples from a mixture which obeys Assumptions 2 an 1. We use the notation S 1 resp. S 2 ) to enote the subspace containing the centers of the istributions in the mixture in View 1 resp. View 2), an notation S 1 resp. S 2 ) to enote the orthogonal complement to the subspace containing the centers of the istributions in the mixture in View 1 resp. View 2). For any matrix A, we use A to enote the L 2 norm or maximum singular value of A. Proofs. Now, we are reay to prove our main theorem. First, we show the following two lemmas, which emonstrate properties of the expecte crosscorrelational matrix across the views. Their proofs are immeiate from Assumptions 2 an 1. Lemma 2 Let v 1 an v 2 be any vectors in S 1 an S 2 respectively. Then, v 1 ) T Σ 12 v 2 > λ min. Lemma 3 Let v 1 resp. v 2 ) be any vector in S 1 resp. S 2 ). Then, for any u 1 V 1 an u 2 V 2, v 1 ) T Σ 12 u 2 = u 1 ) T Σ 12 v 2 = 0. Next, we show that given sufficiently many samples, the subspace spanne by the top k 1 singular vectors of Σ 12 still approximates the subspace containing the means of the istributions comprising the mixture. Finally, we use this fact, along with some results in [AK05] to prove Theorem 1. Our main lemma of this section is the following. Lemma 4 Projection Subspace Lemma) Let v 1 resp. v 2 ) be any vector in S 1 resp. S 2 ). If the number of samples n > c τ 2 λ 2 log2 min wmin τλ minw min ) log 2 1 δ ) for some constant c, then, with probability 1 δ, the length of the projection of v 1 resp. v 2 ) in the subspace spanne by the top k 1 left resp. right) singular vectors of Σ12 is at least 1 τ 2 v 1 resp. 1 τ 2 v 2 ). The main tool in the proof of Lemma 4, is the following lemma, which uses a result ue to [RV07]. Lemma 5 Sample Complexity Lemma) If number of samples n > c ɛ 2 log 2 ) log 2 1 w min ɛw min δ ) the for some constant c, then, with probability at least 1 δ, Σ 12 Σ 12 ɛ, where enotes the L 2 -norm of a matrix. A consequence of Lemma 5 an Lemmas 2 an 3 is the following lemma. Lemma 6 Let n > C ɛ 2 w min log 2 ɛw min ) log 2 1 δ ), for some constant C. Then, with probability 1 δ, the top k 1 singular values of Σ 12 have value at least λ min ɛ. The remaining min 1, 2 ) k + 1 singular values of Σ 12 have value at most ɛ. The proof follows by a combination of Lemmas 2,3, 5 an a triangle inequality. Proof:Of Lemma 5) To prove this lemma, we apply Lemma 7. Observe the block representation of Σ in Equation 1. Moreover, with Σ 11 an Σ 22 in isotropic position, we have that the L 2 norm of Σ 12 is at most 1. Using the triangle inequality, we can write: Σ 12 Σ Σ Σ + Σ 11 Σ 11 + Σ 22 Σ 22 ) where we have applie the triangle inequality to the 2 2 block matrix with off-iagonal entries Σ 12 Σ 12 an with 0 iagonal entries). We now apply Lemma 7 three times, on Σ 11 Σ 11, on Σ 22 Σ 22 an a scale version of Σ Σ. The first two applications follow irectly.

6 For the thir application, we observe that Lemma 7 is rotation invariant, an that scaling each covariance value by some factor s scales the norm of the matrix by at most s. We claim that we can apply Lemma 7 on Σ Σ with s = 4. Since the covariance of any two ranom variables is at most the prouct of their stanar eviations, an since Σ 11 an Σ 22 are I 1 an I 2 respectively, the maximum singular value of Σ 12 is at most 1; the maximum singular value of Σ is therefore at most 4. Our claim follows. The lemma now follows by plugging in n as a function of ɛ, an w min Lemma 7 Let X be a set of n points generate by a mixture of k Gaussians over R, scale such that E[x x T ] = I. If M is the sample covariance matrix of X, then, for n large enough, with probability at least 1 δ, log n log 2n δ M E[M] C ) log1/δ) wmin n where C is a fixe constant, an w min is the minimum mixing weight of any Gaussian in the mixture. Proof: To prove this lemma, we use a concentration result on the L 2 -norms of matrices ue to [RV07]. We observe that each vector x i in the scale space is generate by a Gaussian with some mean µ an maximum irectional variance σ 2. As the total variance of the mixture along any irection is at most 1, w min µ 2 + σ 2 ) 1. Therefore, for all samples x i, with probability at least 1 δ/2, x i µ +σ log 2n δ ). We conition on the fact that the event x i µ + σ log 2n δ ) happens for all i = 1,..., n. The probability of this event is at least 1 δ/2. Conitione on this event, the istributions of the vectors x i are inepenent. Therefore, we can apply Theorem 3.1 in [RV07] on these conitional istributions, to conclue that: Pr[ M E[M] > t] 2e cnt2 /Λ 2 log n where c is a constant, an Λ is an upper boun on the norm of any vector x i. The lemma follows by plugging in t =, an Λ 2 log2n/δ) wmin. Λ 2 log4/δ) log n cn Now we are reay to prove Lemma 4. Proof: Of Lemma 4) For the sake of contraiction, suppose there exists a vector v 1 S 1 such that the projection of v 1 on the top k 1 left singular vectors of Σ 12 is equal to 1 τ 2 v 1, where τ > τ. Then, there exists some unit vector u 1 in V 1 in the orthogonal complement of the space spanne by the top k 1 left singular vectors of Σ 12 such that the projection of v 1 on u 1 is equal to τ v 1. This vector u 1 can be written as: u 1 = τv 1 +1 τ 2 ) 1/2 y 1, where y 1 is in the orthogonal complement of S 1. From Lemma 2, there exists some vector u 2 in S 2, such that v 1 ) Σ 12 u 2 λ min ; from Lemma 3, for this vector u 2, u 1 ) Σ 12 u 2 τλ min. If n > c τ 2 λ 2 log2 min wmin τλ minw min ) log 2 1 δ ), then, from Lemma 6, u 1 ) T Σ12 u 2 τ 2 λ min. Now, since u 1 is in the orthogonal complement of the subspace spanne by the top k 1 left singular vectors of Σ 12, for any vector y 2 in the subspace spanne by the top k 1 right singular vectors of Σ 12, u 1 ) Σ12 y 2 = 0. This, in turn, means that there exists a vector z 2 in V 2 in the orthogonal complement of the subspace spanne by the top k 1 right singular vectors of Σ 12 such that u 1 ) T Σ12 z 2 τ 2 λ min. This implies that the k-th singular value of Σ 12 is at least τ 2 λ min. However, from Lemma 6, all except the top k 1 singular values of Σ 12 are at most τ 3 λ min, which is a contraiction. Finally, we are reay to prove our main theorem. Proof:Of Theorem 1) From Lemma 4, if n > C τ 2 λ 2 log2 min wmin τλ minw min ) log 2 1/δ), then, with probability at least 1 δ, the projection of any vector v in S 1 or S 2 onto the subspace returne by Step 2 of Algorithm 1 has length at least 1 τ 2 v. Therefore, the maximum irectional variance of any D i in any this subspace is at most 1 τ 2 )σ ) 2 + τ 2 σ 2, where σ 2 is the maximum irectional variance of any D i. When τ σ /σ, this is at most 2σ ) 2. From the isotropic conition, σ 1 wmin. Therefore, when σ ) 2 λ 2 min w2 min n > C log 2 σ λ minw min ) log 2 1/δ), the maximum irectional variance of any D i in the mixture in the space output by Step 2 of the Algorithm is at most 2σ ) 2. Since A an B are ranom partitions of the sample set S, the subspace prouce by the action of Step 2 of Algorithm 1 on the set A is inepenent of B, an vice versa. Therefore, when projecte onto the top k 1 SVD subspace of Σ 12 A), the samples from B are istribute as a mixture of k 1)-imensional Gaussians. The theorem follows from the bouns in the previous paragraph, an Theorem 1 of [AK05].

7 5. Experiments 5.1. Auio-visual speaker clustering In the first set of experiments, we consier clustering either auio or face images of speakers. We use a subset of the ViTIMIT atabase [?] consisting of 41 speakers, speaking 10 sentences about secons) each, recore at 25 frames per secon in a stuio environment with no significant lighting or pose variation. The auio features are stanar 12-imensional mel cepstra [?] compute every 10ms over a ms winow, an finally concatenate over a winow of 0ms before an after the current frame, for a total of 1584 imensions. The vieo features are pixels of the face region extracte from each image 2394 imensions). We consier the target cluster variable to be the speaker. We use either CCA or PCA to project the ata to some lower imensionality N. In the case of CCA, we initially project the ata to an intermeiate imensionality M using PCA to reuce the effects of spurious correlations. For the results reporte here, N is typically an M is typically 100 for images an 1000 for auio. The parameters were selecte using a hel-out set. For CCA, we ranomize the vectors of one view in each sentence, to reuce correlations between the views ue to certain other latent variables such as the current phoneme. We then cluster either view using k-means into 82 clusters 2 per speaker). To alleviate the problem of local minima foun by k- means, each clustering consists of 5 runs of k-means, an the one with the lowest k-means score is taken as the final cluster assignment. Similarly to [?], we measure clustering performance using the conitional entropy of the speaker s given the cluster c, Hs c). We report the results in terms of conitional perplexity, 2 Hs c), which is the mean number of speakers corresponing to each cluster. Table 1 shows results on the raw ata, as well as with synthetic occlusions an translations of the image ata. Consiering the extremely clean visual environment, we expect PCA to o very well on the image ata. Inee, PCA provies an almost perfect clustering of the raw images an CCA oes not improve it. However, CCA far outperforms PCA when clustering the more challenging auio view. When synthetic occlusions or translations are applie to the images, the performance of PCA-base clustering is greatly egrae. CCA is unaffecte in the case of occlusion; in the case of translation, CCA-base image clustering is egrae similarly to PCA, but auio clustering is almost unaffecte. In other wors, even when the image ata are egrae, CCA is able to recover a goo clustering in at least one of the views. For a more e- PCA CCA Images Auio Images + occlusion Auio + occlusion Images + translation Auio + translation Table 1. Conitional perplexities of the speaker given the cluster, using PCA or CCA bases. + occlusion an + translation inicate that the images are corrupte with occlusion/translation; the auio is unchange, however. taile look at the clustering behavior, Figures 1a-) show the istributions of clusters for each speaker in several conitions Clustering Wikipeia articles Next we consier the task of clustering Wikipeia articles, base on either their text or their incoming an outgoing links. The link structure L is represente as a concatenation of to an from link incience vectors, where each element Li) is the number of times the current article links to/from article i. The article text is represente as a bag-of-wors feature vector, i.e. the raw count of each wor in the article. A lexicon of about 8 million wors an a list of about 12 million articles were use to construct the two feature vectors. Since the imensionality of the feature vectors is very high over million for the link view), we use ranom projection to reuce the imensionality to a computationally manageable level. We present clustering experiments on a subset of Wikipeia consisting of 128,327 articles. We use either PCA or CCA to reuce the feature vectors to the final esire imensionality, followe by clustering. In these experiments, we use a hierarchical clustering proceure, as a flat clustering is rather poor with either PCA or CCA CCA still usually outperforms PCA, however). In the hierarchical proceure, all points are initially consiere to be in a single cluster. Next, we iteratively pick the largest cluster, reuce the imensionality using PCA or CCA on the points in this cluster, an use k-means to break the cluster into smaller sub-clusters for some fixe k), until we reach the total esire number of clusters. The intuition for this is that ifferent clusters may have ifferent natural subspaces. As before, we evaluate the clustering using the conitional perplexity of the article category a as given by Wikipeia) given the cluster c, 2 Ha c). For each arti-

8 speaker a) AV: Auio, PCA basis speaker c) AV: Images + occlusion, PCA basis perplexity e) Wikipeia: Category perplexity hierarchical CCA hierarchical PCA cluster cluster number of clusters 5 10 b) AV: Auio, CCA basis 5 10 ) AV: Images + occlusion, CCA basis f) Wikipeia: Cluster perplexity balance clustering hierarchical CCA hierarchical PCA speaker speaker Entropy cluster cluster number of clusters Figure 1. a-) Distributions of cluster assignments per speaker in auio-visual experiments. The color of each cell s, c) correspons to the empirical probability pc s) arker = higher). e-f) Wikipeia experiments: e) Conitional perplexity of article category given cluster, 2 Ha c), as a function of the number of clusters. f) Perplexity of the cluster istribution 2 Hc) ) as a function of the number of clusters. cle we use the first category liste in the article. The 128,327 articles inclue roughly 15,000 categories, of which we use the 500 most frequent ones, which cover 73,145 articles. While the clustering is performe on all 128,327 articles, the reporte entropies are for the 73,145 articles covere by the 500 categories. Each sub-clustering consists of 10 runs of k-means, an the one with the lowest k-means score is taken as the final cluster assignment. Figure 1e) shows the conitional perplexity versus the number of clusters for PCA- an CCA-base hierarchical clustering. For any number of clusters, CCA prouces better clusterings, i.e. ones with lower perplexity. Note that the reuction in entropy with larger number of clusters is expecte for any approach; the relevant point is the ifference for any given number of clusters.) In aition, the tree structures of the PCA-/CCA-base clusterings are qualitatively ifferent. With PCA-base clustering, most points are assigne to a few large clusters, with the remaining clusters consisting of only a few points. CCA-base hierarchical clustering prouces more balance clusters. To see this, in Figure 1f) we show the perplexity of the cluster istribution, 2 Hc), versus the number of clusters. For about 25 or more clusters, the CCAbase clusterings have higher perplexity, inicating a more uniform istribution of clusters than in PCAbase clustering. References [AK05] [AM05] [AZ07] S. Arora an R. Kannan. Learning mixtures of separate nonspherical gaussians. Annals of Applie Probability, 151A):69 92, 05. D. Achlioptas an F. McSherry. On spectral learning of mixtures of istributions. In Proceeings of the 18th Annual Conference on Learning Theory, pages , 05. Rie Kubota Ano an Tong Zhang. Two-view feature generation moel for semi-supervise learning. In ICML 07: Proceeings of the 24th international conference on Machine learning, pages 25 32, New York, NY, USA, 07. ACM. [BM98] Avrim Blum an Tom Mitchell. Combining labele an unlabele ata with co-training. In COLT: Proceeings of the Workshop on Computational Learning Theory, Morgan Kaufmann Publishers, pages , [BV08] S. C. Brubaker an S. Vempala. Isotropic pca an affineinvariant clustering. In Proc. of Founations of Computer Science, 08. [CR08] K. Chauhuri an S. Rao. Learning mixtures of istributions using correlations an inepenence. In In Proc. of Conference on Learning Theory, 08. [Das99] S. Dasgupta. Learning mixtures of gaussians. In Proceeings of the th IEEE Symposium on Founations of Computer S cience, pages , [DS00] S. Dasgupta an L. Schulman. A two-roun variant of em for gaussian mixtures. In Sixteenth Conference on Uncertainty in Artificial Intelligence UAI), 00. [KF07] Sham M. Kakae an Dean P. Foster. Multi-view regression via canonical correlation analysis. In Naer H. Bshouty an Clauio Gentile, eitors, COLT, volume 4539 of Lecture Notes in Computer Science, pages Springer, 07.

9 [KSV05] [RV07] [VW02] R. Kannan, H. Salmasian, an S. Vempala. The spectral metho for general mixture moels. In Proceeings of the 18th Annual Conference on Learning Theory, 05. M. Ruelson an R. Vershynin. Sampling from large matrices: An approach through geometric functional analysis. Journal of the ACM, 07. V. Vempala an G. Wang. A spectral algorithm of learning mixtures of istributions. In Proceeings of the 43r IEEE Symposium on Founations of Computer Science, pages , 02.

Multi-View Clustering via Canonical Correlation Analysis

Multi-View Clustering via Canonical Correlation Analysis Technical Report TTI-TR-2008-5 Multi-View Clustering via Canonical Correlation Analysis Kamalika Chauhuri UC San Diego Sham M. Kakae Toyota Technological Institute at Chicago ABSTRACT Clustering ata in

More information

Multi-View Clustering via Canonical Correlation Analysis

Multi-View Clustering via Canonical Correlation Analysis Kamalika Chauhuri ITA, UC San Diego, 9500 Gilman Drive, La Jolla, CA Sham M. Kakae Karen Livescu Karthik Sriharan Toyota Technological Institute at Chicago, 6045 S. Kenwoo Ave., Chicago, IL kamalika@soe.ucs.eu

More information

Multi-View Clustering via Canonical Correlation Analysis

Multi-View Clustering via Canonical Correlation Analysis Kamalika Chauhuri ITA, UC San Diego, 9500 Gilman Drive, La Jolla, CA Sham M. Kakae Karen Livescu Karthik Sriharan Toyota Technological Institute at Chicago, 6045 S. Kenwoo Ave., Chicago, IL kamalika@soe.ucs.eu

More information

Lecture Introduction. 2 Examples of Measure Concentration. 3 The Johnson-Lindenstrauss Lemma. CS-621 Theory Gems November 28, 2012

Lecture Introduction. 2 Examples of Measure Concentration. 3 The Johnson-Lindenstrauss Lemma. CS-621 Theory Gems November 28, 2012 CS-6 Theory Gems November 8, 0 Lecture Lecturer: Alesaner Mąry Scribes: Alhussein Fawzi, Dorina Thanou Introuction Toay, we will briefly iscuss an important technique in probability theory measure concentration

More information

u!i = a T u = 0. Then S satisfies

u!i = a T u = 0. Then S satisfies Deterministic Conitions for Subspace Ientifiability from Incomplete Sampling Daniel L Pimentel-Alarcón, Nigel Boston, Robert D Nowak University of Wisconsin-Maison Abstract Consier an r-imensional subspace

More information

A PAC-Bayesian Approach to Spectrally-Normalized Margin Bounds for Neural Networks

A PAC-Bayesian Approach to Spectrally-Normalized Margin Bounds for Neural Networks A PAC-Bayesian Approach to Spectrally-Normalize Margin Bouns for Neural Networks Behnam Neyshabur, Srinah Bhojanapalli, Davi McAllester, Nathan Srebro Toyota Technological Institute at Chicago {bneyshabur,

More information

Analyzing Tensor Power Method Dynamics in Overcomplete Regime

Analyzing Tensor Power Method Dynamics in Overcomplete Regime Journal of Machine Learning Research 18 (2017) 1-40 Submitte 9/15; Revise 11/16; Publishe 4/17 Analyzing Tensor Power Metho Dynamics in Overcomplete Regime Animashree Ananumar Department of Electrical

More information

Survey Sampling. 1 Design-based Inference. Kosuke Imai Department of Politics, Princeton University. February 19, 2013

Survey Sampling. 1 Design-based Inference. Kosuke Imai Department of Politics, Princeton University. February 19, 2013 Survey Sampling Kosuke Imai Department of Politics, Princeton University February 19, 2013 Survey sampling is one of the most commonly use ata collection methos for social scientists. We begin by escribing

More information

Lower bounds on Locality Sensitive Hashing

Lower bounds on Locality Sensitive Hashing Lower bouns on Locality Sensitive Hashing Rajeev Motwani Assaf Naor Rina Panigrahy Abstract Given a metric space (X, X ), c 1, r > 0, an p, q [0, 1], a istribution over mappings H : X N is calle a (r,

More information

Permanent vs. Determinant

Permanent vs. Determinant Permanent vs. Determinant Frank Ban Introuction A major problem in theoretical computer science is the Permanent vs. Determinant problem. It asks: given an n by n matrix of ineterminates A = (a i,j ) an

More information

Influence of weight initialization on multilayer perceptron performance

Influence of weight initialization on multilayer perceptron performance Influence of weight initialization on multilayer perceptron performance M. Karouia (1,2) T. Denœux (1) R. Lengellé (1) (1) Université e Compiègne U.R.A. CNRS 817 Heuiasyc BP 649 - F-66 Compiègne ceex -

More information

Necessary and Sufficient Conditions for Sketched Subspace Clustering

Necessary and Sufficient Conditions for Sketched Subspace Clustering Necessary an Sufficient Conitions for Sketche Subspace Clustering Daniel Pimentel-Alarcón, Laura Balzano 2, Robert Nowak University of Wisconsin-Maison, 2 University of Michigan-Ann Arbor Abstract This

More information

This module is part of the. Memobust Handbook. on Methodology of Modern Business Statistics

This module is part of the. Memobust Handbook. on Methodology of Modern Business Statistics This moule is part of the Memobust Hanbook on Methoology of Moern Business Statistics 26 March 2014 Metho: Balance Sampling for Multi-Way Stratification Contents General section... 3 1. Summary... 3 2.

More information

Least-Squares Regression on Sparse Spaces

Least-Squares Regression on Sparse Spaces Least-Squares Regression on Sparse Spaces Yuri Grinberg, Mahi Milani Far, Joelle Pineau School of Computer Science McGill University Montreal, Canaa {ygrinb,mmilan1,jpineau}@cs.mcgill.ca 1 Introuction

More information

Lecture 6 : Dimensionality Reduction

Lecture 6 : Dimensionality Reduction CPS290: Algorithmic Founations of Data Science February 3, 207 Lecture 6 : Dimensionality Reuction Lecturer: Kamesh Munagala Scribe: Kamesh Munagala In this lecture, we will consier the roblem of maing

More information

Robust Forward Algorithms via PAC-Bayes and Laplace Distributions. ω Q. Pr (y(ω x) < 0) = Pr A k

Robust Forward Algorithms via PAC-Bayes and Laplace Distributions. ω Q. Pr (y(ω x) < 0) = Pr A k A Proof of Lemma 2 B Proof of Lemma 3 Proof: Since the support of LL istributions is R, two such istributions are equivalent absolutely continuous with respect to each other an the ivergence is well-efine

More information

LATTICE-BASED D-OPTIMUM DESIGN FOR FOURIER REGRESSION

LATTICE-BASED D-OPTIMUM DESIGN FOR FOURIER REGRESSION The Annals of Statistics 1997, Vol. 25, No. 6, 2313 2327 LATTICE-BASED D-OPTIMUM DESIGN FOR FOURIER REGRESSION By Eva Riccomagno, 1 Rainer Schwabe 2 an Henry P. Wynn 1 University of Warwick, Technische

More information

Database-friendly Random Projections

Database-friendly Random Projections Database-frienly Ranom Projections Dimitris Achlioptas Microsoft ABSTRACT A classic result of Johnson an Linenstrauss asserts that any set of n points in -imensional Eucliean space can be embee into k-imensional

More information

Acute sets in Euclidean spaces

Acute sets in Euclidean spaces Acute sets in Eucliean spaces Viktor Harangi April, 011 Abstract A finite set H in R is calle an acute set if any angle etermine by three points of H is acute. We examine the maximal carinality α() of

More information

Time-of-Arrival Estimation in Non-Line-Of-Sight Environments

Time-of-Arrival Estimation in Non-Line-Of-Sight Environments 2 Conference on Information Sciences an Systems, The Johns Hopkins University, March 2, 2 Time-of-Arrival Estimation in Non-Line-Of-Sight Environments Sinan Gezici, Hisashi Kobayashi an H. Vincent Poor

More information

Robust Low Rank Kernel Embeddings of Multivariate Distributions

Robust Low Rank Kernel Embeddings of Multivariate Distributions Robust Low Rank Kernel Embeings of Multivariate Distributions Le Song, Bo Dai College of Computing, Georgia Institute of Technology lsong@cc.gatech.eu, boai@gatech.eu Abstract Kernel embeing of istributions

More information

Multi-View Dimensionality Reduction via Canonical Correlation Analysis

Multi-View Dimensionality Reduction via Canonical Correlation Analysis Technical Report TTI-TR-2008-4 Multi-View Dimensionality Reduction via Canonical Correlation Analysis Dean P. Foster University of Pennsylvania Sham M. Kakade Toyota Technological Institute at Chicago

More information

7.1 Support Vector Machine

7.1 Support Vector Machine 67577 Intro. to Machine Learning Fall semester, 006/7 Lecture 7: Support Vector Machines an Kernel Functions II Lecturer: Amnon Shashua Scribe: Amnon Shashua 7. Support Vector Machine We return now to

More information

Estimation of the Maximum Domination Value in Multi-Dimensional Data Sets

Estimation of the Maximum Domination Value in Multi-Dimensional Data Sets Proceeings of the 4th East-European Conference on Avances in Databases an Information Systems ADBIS) 200 Estimation of the Maximum Domination Value in Multi-Dimensional Data Sets Eleftherios Tiakas, Apostolos.

More information

The total derivative. Chapter Lagrangian and Eulerian approaches

The total derivative. Chapter Lagrangian and Eulerian approaches Chapter 5 The total erivative 51 Lagrangian an Eulerian approaches The representation of a flui through scalar or vector fiels means that each physical quantity uner consieration is escribe as a function

More information

Agmon Kolmogorov Inequalities on l 2 (Z d )

Agmon Kolmogorov Inequalities on l 2 (Z d ) Journal of Mathematics Research; Vol. 6, No. ; 04 ISSN 96-9795 E-ISSN 96-9809 Publishe by Canaian Center of Science an Eucation Agmon Kolmogorov Inequalities on l (Z ) Arman Sahovic Mathematics Department,

More information

Capacity Analysis of MIMO Systems with Unknown Channel State Information

Capacity Analysis of MIMO Systems with Unknown Channel State Information Capacity Analysis of MIMO Systems with Unknown Channel State Information Jun Zheng an Bhaskar D. Rao Dept. of Electrical an Computer Engineering University of California at San Diego e-mail: juzheng@ucs.eu,

More information

Convergence of Random Walks

Convergence of Random Walks Chapter 16 Convergence of Ranom Walks This lecture examines the convergence of ranom walks to the Wiener process. This is very important both physically an statistically, an illustrates the utility of

More information

An Optimal Algorithm for Bandit and Zero-Order Convex Optimization with Two-Point Feedback

An Optimal Algorithm for Bandit and Zero-Order Convex Optimization with Two-Point Feedback Journal of Machine Learning Research 8 07) - Submitte /6; Publishe 5/7 An Optimal Algorithm for Banit an Zero-Orer Convex Optimization with wo-point Feeback Oha Shamir Department of Computer Science an

More information

THE EFFICIENCIES OF THE SPATIAL MEDIAN AND SPATIAL SIGN COVARIANCE MATRIX FOR ELLIPTICALLY SYMMETRIC DISTRIBUTIONS

THE EFFICIENCIES OF THE SPATIAL MEDIAN AND SPATIAL SIGN COVARIANCE MATRIX FOR ELLIPTICALLY SYMMETRIC DISTRIBUTIONS THE EFFICIENCIES OF THE SPATIAL MEDIAN AND SPATIAL SIGN COVARIANCE MATRIX FOR ELLIPTICALLY SYMMETRIC DISTRIBUTIONS BY ANDREW F. MAGYAR A issertation submitte to the Grauate School New Brunswick Rutgers,

More information

Computing Exact Confidence Coefficients of Simultaneous Confidence Intervals for Multinomial Proportions and their Functions

Computing Exact Confidence Coefficients of Simultaneous Confidence Intervals for Multinomial Proportions and their Functions Working Paper 2013:5 Department of Statistics Computing Exact Confience Coefficients of Simultaneous Confience Intervals for Multinomial Proportions an their Functions Shaobo Jin Working Paper 2013:5

More information

Diagonalization of Matrices Dr. E. Jacobs

Diagonalization of Matrices Dr. E. Jacobs Diagonalization of Matrices Dr. E. Jacobs One of the very interesting lessons in this course is how certain algebraic techniques can be use to solve ifferential equations. The purpose of these notes is

More information

arxiv: v1 [cs.lg] 22 Mar 2014

arxiv: v1 [cs.lg] 22 Mar 2014 CUR lgorithm with Incomplete Matrix Observation Rong Jin an Shenghuo Zhu Dept. of Computer Science an Engineering, Michigan State University, rongjin@msu.eu NEC Laboratories merica, Inc., zsh@nec-labs.com

More information

Problem Sheet 2: Eigenvalues and eigenvectors and their use in solving linear ODEs

Problem Sheet 2: Eigenvalues and eigenvectors and their use in solving linear ODEs Problem Sheet 2: Eigenvalues an eigenvectors an their use in solving linear ODEs If you fin any typos/errors in this problem sheet please email jk28@icacuk The material in this problem sheet is not examinable

More information

On combinatorial approaches to compressed sensing

On combinatorial approaches to compressed sensing On combinatorial approaches to compresse sensing Abolreza Abolhosseini Moghaam an Hayer Raha Department of Electrical an Computer Engineering, Michigan State University, East Lansing, MI, U.S. Emails:{abolhos,raha}@msu.eu

More information

A Unified Theorem on SDP Rank Reduction

A Unified Theorem on SDP Rank Reduction A Unifie heorem on SDP Ran Reuction Anthony Man Cho So, Yinyu Ye, Jiawei Zhang November 9, 006 Abstract We consier the problem of fining a low ran approximate solution to a system of linear equations in

More information

Parameter estimation: A new approach to weighting a priori information

Parameter estimation: A new approach to weighting a priori information Parameter estimation: A new approach to weighting a priori information J.L. Mea Department of Mathematics, Boise State University, Boise, ID 83725-555 E-mail: jmea@boisestate.eu Abstract. We propose a

More information

Linear First-Order Equations

Linear First-Order Equations 5 Linear First-Orer Equations Linear first-orer ifferential equations make up another important class of ifferential equations that commonly arise in applications an are relatively easy to solve (in theory)

More information

A simple model for the small-strain behaviour of soils

A simple model for the small-strain behaviour of soils A simple moel for the small-strain behaviour of soils José Jorge Naer Department of Structural an Geotechnical ngineering, Polytechnic School, University of São Paulo 05508-900, São Paulo, Brazil, e-mail:

More information

Proof of SPNs as Mixture of Trees

Proof of SPNs as Mixture of Trees A Proof of SPNs as Mixture of Trees Theorem 1. If T is an inuce SPN from a complete an ecomposable SPN S, then T is a tree that is complete an ecomposable. Proof. Argue by contraiction that T is not a

More information

Flexible High-Dimensional Classification Machines and Their Asymptotic Properties

Flexible High-Dimensional Classification Machines and Their Asymptotic Properties Journal of Machine Learning Research 16 (2015) 1547-1572 Submitte 1/14; Revise 9/14; Publishe 8/15 Flexible High-Dimensional Classification Machines an Their Asymptotic Properties Xingye Qiao Department

More information

Topic 7: Convergence of Random Variables

Topic 7: Convergence of Random Variables Topic 7: Convergence of Ranom Variables Course 003, 2016 Page 0 The Inference Problem So far, our starting point has been a given probability space (S, F, P). We now look at how to generate information

More information

LECTURE NOTES ON DVORETZKY S THEOREM

LECTURE NOTES ON DVORETZKY S THEOREM LECTURE NOTES ON DVORETZKY S THEOREM STEVEN HEILMAN Abstract. We present the first half of the paper [S]. In particular, the results below, unless otherwise state, shoul be attribute to G. Schechtman.

More information

Lectures - Week 10 Introduction to Ordinary Differential Equations (ODES) First Order Linear ODEs

Lectures - Week 10 Introduction to Ordinary Differential Equations (ODES) First Order Linear ODEs Lectures - Week 10 Introuction to Orinary Differential Equations (ODES) First Orer Linear ODEs When stuying ODEs we are consiering functions of one inepenent variable, e.g., f(x), where x is the inepenent

More information

Speaker Adaptation Based on Sparse and Low-rank Eigenphone Matrix Estimation

Speaker Adaptation Based on Sparse and Low-rank Eigenphone Matrix Estimation INTERSPEECH 2014 Speaker Aaptation Base on Sparse an Low-rank Eigenphone Matrix Estimation Wen-Lin Zhang 1, Dan Qu 1, Wei-Qiang Zhang 2, Bi-Cheng Li 1 1 Zhengzhou Information Science an Technology Institute,

More information

Lecture 5. Symmetric Shearer s Lemma

Lecture 5. Symmetric Shearer s Lemma Stanfor University Spring 208 Math 233: Non-constructive methos in combinatorics Instructor: Jan Vonrák Lecture ate: January 23, 208 Original scribe: Erik Bates Lecture 5 Symmetric Shearer s Lemma Here

More information

CUSTOMER REVIEW FEATURE EXTRACTION Heng Ren, Jingye Wang, and Tony Wu

CUSTOMER REVIEW FEATURE EXTRACTION Heng Ren, Jingye Wang, and Tony Wu CUSTOMER REVIEW FEATURE EXTRACTION Heng Ren, Jingye Wang, an Tony Wu Abstract Popular proucts often have thousans of reviews that contain far too much information for customers to igest. Our goal for the

More information

FLUCTUATIONS IN THE NUMBER OF POINTS ON SMOOTH PLANE CURVES OVER FINITE FIELDS. 1. Introduction

FLUCTUATIONS IN THE NUMBER OF POINTS ON SMOOTH PLANE CURVES OVER FINITE FIELDS. 1. Introduction FLUCTUATIONS IN THE NUMBER OF POINTS ON SMOOTH PLANE CURVES OVER FINITE FIELDS ALINA BUCUR, CHANTAL DAVID, BROOKE FEIGON, MATILDE LALÍN 1 Introuction In this note, we stuy the fluctuations in the number

More information

Logarithmic spurious regressions

Logarithmic spurious regressions Logarithmic spurious regressions Robert M. e Jong Michigan State University February 5, 22 Abstract Spurious regressions, i.e. regressions in which an integrate process is regresse on another integrate

More information

Lecture 2: Correlated Topic Model

Lecture 2: Correlated Topic Model Probabilistic Moels for Unsupervise Learning Spring 203 Lecture 2: Correlate Topic Moel Inference for Correlate Topic Moel Yuan Yuan First of all, let us make some claims about the parameters an variables

More information

Lecture 6: Calculus. In Song Kim. September 7, 2011

Lecture 6: Calculus. In Song Kim. September 7, 2011 Lecture 6: Calculus In Song Kim September 7, 20 Introuction to Differential Calculus In our previous lecture we came up with several ways to analyze functions. We saw previously that the slope of a linear

More information

A note on asymptotic formulae for one-dimensional network flow problems Carlos F. Daganzo and Karen R. Smilowitz

A note on asymptotic formulae for one-dimensional network flow problems Carlos F. Daganzo and Karen R. Smilowitz A note on asymptotic formulae for one-imensional network flow problems Carlos F. Daganzo an Karen R. Smilowitz (to appear in Annals of Operations Research) Abstract This note evelops asymptotic formulae

More information

Improving Estimation Accuracy in Nonrandomized Response Questioning Methods by Multiple Answers

Improving Estimation Accuracy in Nonrandomized Response Questioning Methods by Multiple Answers International Journal of Statistics an Probability; Vol 6, No 5; September 207 ISSN 927-7032 E-ISSN 927-7040 Publishe by Canaian Center of Science an Eucation Improving Estimation Accuracy in Nonranomize

More information

19 Eigenvalues, Eigenvectors, Ordinary Differential Equations, and Control

19 Eigenvalues, Eigenvectors, Ordinary Differential Equations, and Control 19 Eigenvalues, Eigenvectors, Orinary Differential Equations, an Control This section introuces eigenvalues an eigenvectors of a matrix, an iscusses the role of the eigenvalues in etermining the behavior

More information

A Course in Machine Learning

A Course in Machine Learning A Course in Machine Learning Hal Daumé III 12 EFFICIENT LEARNING So far, our focus has been on moels of learning an basic algorithms for those moels. We have not place much emphasis on how to learn quickly.

More information

Learning Mixtures of Gaussians with Maximum-a-posteriori Oracle

Learning Mixtures of Gaussians with Maximum-a-posteriori Oracle Satyaki Mahalanabis Dept of Computer Science University of Rochester smahalan@csrochestereu Abstract We consier the problem of estimating the parameters of a mixture of istributions, where each component

More information

d-dimensional Arrangement Revisited

d-dimensional Arrangement Revisited -Dimensional Arrangement Revisite Daniel Rotter Jens Vygen Research Institute for Discrete Mathematics University of Bonn Revise version: April 5, 013 Abstract We revisit the -imensional arrangement problem

More information

1 dx. where is a large constant, i.e., 1, (7.6) and Px is of the order of unity. Indeed, if px is given by (7.5), the inequality (7.

1 dx. where is a large constant, i.e., 1, (7.6) and Px is of the order of unity. Indeed, if px is given by (7.5), the inequality (7. Lectures Nine an Ten The WKB Approximation The WKB metho is a powerful tool to obtain solutions for many physical problems It is generally applicable to problems of wave propagation in which the frequency

More information

Math 1B, lecture 8: Integration by parts

Math 1B, lecture 8: Integration by parts Math B, lecture 8: Integration by parts Nathan Pflueger 23 September 2 Introuction Integration by parts, similarly to integration by substitution, reverses a well-known technique of ifferentiation an explores

More information

Lower Bounds for the Smoothed Number of Pareto optimal Solutions

Lower Bounds for the Smoothed Number of Pareto optimal Solutions Lower Bouns for the Smoothe Number of Pareto optimal Solutions Tobias Brunsch an Heiko Röglin Department of Computer Science, University of Bonn, Germany brunsch@cs.uni-bonn.e, heiko@roeglin.org Abstract.

More information

Expected Value of Partial Perfect Information

Expected Value of Partial Perfect Information Expecte Value of Partial Perfect Information Mike Giles 1, Takashi Goa 2, Howar Thom 3 Wei Fang 1, Zhenru Wang 1 1 Mathematical Institute, University of Oxfor 2 School of Engineering, University of Tokyo

More information

On the Surprising Behavior of Distance Metrics in High Dimensional Space

On the Surprising Behavior of Distance Metrics in High Dimensional Space On the Surprising Behavior of Distance Metrics in High Dimensional Space Charu C. Aggarwal, Alexaner Hinneburg 2, an Daniel A. Keim 2 IBM T. J. Watson Research Center Yortown Heights, NY 0598, USA. charu@watson.ibm.com

More information

ELEC3114 Control Systems 1

ELEC3114 Control Systems 1 ELEC34 Control Systems Linear Systems - Moelling - Some Issues Session 2, 2007 Introuction Linear systems may be represente in a number of ifferent ways. Figure shows the relationship between various representations.

More information

Robustness and Perturbations of Minimal Bases

Robustness and Perturbations of Minimal Bases Robustness an Perturbations of Minimal Bases Paul Van Dooren an Froilán M Dopico December 9, 2016 Abstract Polynomial minimal bases of rational vector subspaces are a classical concept that plays an important

More information

Collapsed Gibbs and Variational Methods for LDA. Example Collapsed MoG Sampling

Collapsed Gibbs and Variational Methods for LDA. Example Collapsed MoG Sampling Case Stuy : Document Retrieval Collapse Gibbs an Variational Methos for LDA Machine Learning/Statistics for Big Data CSE599C/STAT59, University of Washington Emily Fox 0 Emily Fox February 7 th, 0 Example

More information

Tractability results for weighted Banach spaces of smooth functions

Tractability results for weighted Banach spaces of smooth functions Tractability results for weighte Banach spaces of smooth functions Markus Weimar Mathematisches Institut, Universität Jena Ernst-Abbe-Platz 2, 07740 Jena, Germany email: markus.weimar@uni-jena.e March

More information

SYSTEMS OF DIFFERENTIAL EQUATIONS, EULER S FORMULA. where L is some constant, usually called the Lipschitz constant. An example is

SYSTEMS OF DIFFERENTIAL EQUATIONS, EULER S FORMULA. where L is some constant, usually called the Lipschitz constant. An example is SYSTEMS OF DIFFERENTIAL EQUATIONS, EULER S FORMULA. Uniqueness for solutions of ifferential equations. We consier the system of ifferential equations given by x = v( x), () t with a given initial conition

More information

Local Linear ICA for Mutual Information Estimation in Feature Selection

Local Linear ICA for Mutual Information Estimation in Feature Selection Local Linear ICA for Mutual Information Estimation in Feature Selection Tian Lan, Deniz Erogmus Department of Biomeical Engineering, OGI, Oregon Health & Science University, Portlan, Oregon, USA E-mail:

More information

arxiv: v1 [math.mg] 10 Apr 2018

arxiv: v1 [math.mg] 10 Apr 2018 ON THE VOLUME BOUND IN THE DVORETZKY ROGERS LEMMA FERENC FODOR, MÁRTON NASZÓDI, AND TAMÁS ZARNÓCZ arxiv:1804.03444v1 [math.mg] 10 Apr 2018 Abstract. The classical Dvoretzky Rogers lemma provies a eterministic

More information

PDE Notes, Lecture #11

PDE Notes, Lecture #11 PDE Notes, Lecture # from Professor Jalal Shatah s Lectures Febuary 9th, 2009 Sobolev Spaces Recall that for u L loc we can efine the weak erivative Du by Du, φ := udφ φ C0 If v L loc such that Du, φ =

More information

Thermal conductivity of graded composites: Numerical simulations and an effective medium approximation

Thermal conductivity of graded composites: Numerical simulations and an effective medium approximation JOURNAL OF MATERIALS SCIENCE 34 (999)5497 5503 Thermal conuctivity of grae composites: Numerical simulations an an effective meium approximation P. M. HUI Department of Physics, The Chinese University

More information

The canonical controllers and regular interconnection

The canonical controllers and regular interconnection Systems & Control Letters ( www.elsevier.com/locate/sysconle The canonical controllers an regular interconnection A.A. Julius a,, J.C. Willems b, M.N. Belur c, H.L. Trentelman a Department of Applie Mathematics,

More information

Designing of Acceptance Double Sampling Plan for Life Test Based on Percentiles of Exponentiated Rayleigh Distribution

Designing of Acceptance Double Sampling Plan for Life Test Based on Percentiles of Exponentiated Rayleigh Distribution International Journal of Statistics an Systems ISSN 973-675 Volume, Number 3 (7), pp. 475-484 Research Inia Publications http://www.ripublication.com Designing of Acceptance Double Sampling Plan for Life

More information

SYNCHRONOUS SEQUENTIAL CIRCUITS

SYNCHRONOUS SEQUENTIAL CIRCUITS CHAPTER SYNCHRONOUS SEUENTIAL CIRCUITS Registers an counters, two very common synchronous sequential circuits, are introuce in this chapter. Register is a igital circuit for storing information. Contents

More information

Homework 2 EM, Mixture Models, PCA, Dualitys

Homework 2 EM, Mixture Models, PCA, Dualitys Homework 2 EM, Mixture Moels, PCA, Dualitys CMU 10-715: Machine Learning (Fall 2015) http://www.cs.cmu.eu/~bapoczos/classes/ml10715_2015fall/ OUT: Oct 5, 2015 DUE: Oct 19, 2015, 10:20 AM Guielines The

More information

Separation of Variables

Separation of Variables Physics 342 Lecture 1 Separation of Variables Lecture 1 Physics 342 Quantum Mechanics I Monay, January 25th, 2010 There are three basic mathematical tools we nee, an then we can begin working on the physical

More information

arxiv: v4 [math.pr] 27 Jul 2016

arxiv: v4 [math.pr] 27 Jul 2016 The Asymptotic Distribution of the Determinant of a Ranom Correlation Matrix arxiv:309768v4 mathpr] 7 Jul 06 AM Hanea a, & GF Nane b a Centre of xcellence for Biosecurity Risk Analysis, University of Melbourne,

More information

PLAL: Cluster-based Active Learning

PLAL: Cluster-based Active Learning JMLR: Workshop an Conference Proceeings vol 3 (13) 1 22 PLAL: Cluster-base Active Learning Ruth Urner rurner@cs.uwaterloo.ca School of Computer Science, University of Waterloo, Canaa, ON, N2L 3G1 Sharon

More information

A Sketch of Menshikov s Theorem

A Sketch of Menshikov s Theorem A Sketch of Menshikov s Theorem Thomas Bao March 14, 2010 Abstract Let Λ be an infinite, locally finite oriente multi-graph with C Λ finite an strongly connecte, an let p

More information

TEMPORAL AND TIME-FREQUENCY CORRELATION-BASED BLIND SOURCE SEPARATION METHODS. Yannick DEVILLE

TEMPORAL AND TIME-FREQUENCY CORRELATION-BASED BLIND SOURCE SEPARATION METHODS. Yannick DEVILLE TEMPORAL AND TIME-FREQUENCY CORRELATION-BASED BLIND SOURCE SEPARATION METHODS Yannick DEVILLE Université Paul Sabatier Laboratoire Acoustique, Métrologie, Instrumentation Bât. 3RB2, 8 Route e Narbonne,

More information

THE GENUINE OMEGA-REGULAR UNITARY DUAL OF THE METAPLECTIC GROUP

THE GENUINE OMEGA-REGULAR UNITARY DUAL OF THE METAPLECTIC GROUP THE GENUINE OMEGA-REGULAR UNITARY DUAL OF THE METAPLECTIC GROUP ALESSANDRA PANTANO, ANNEGRET PAUL, AND SUSANA A. SALAMANCA-RIBA Abstract. We classify all genuine unitary representations of the metaplectic

More information

Homework 2 Solutions EM, Mixture Models, PCA, Dualitys

Homework 2 Solutions EM, Mixture Models, PCA, Dualitys Homewor Solutions EM, Mixture Moels, PCA, Dualitys CMU 0-75: Machine Learning Fall 05 http://www.cs.cmu.eu/~bapoczos/classes/ml075_05fall/ OUT: Oct 5, 05 DUE: Oct 9, 05, 0:0 AM An EM algorithm for a Mixture

More information

arxiv: v4 [cs.ds] 7 Mar 2014

arxiv: v4 [cs.ds] 7 Mar 2014 Analysis of Agglomerative Clustering Marcel R. Ackermann Johannes Blömer Daniel Kuntze Christian Sohler arxiv:101.697v [cs.ds] 7 Mar 01 Abstract The iameter k-clustering problem is the problem of partitioning

More information

Image Denoising Using Spatial Adaptive Thresholding

Image Denoising Using Spatial Adaptive Thresholding International Journal of Engineering Technology, Management an Applie Sciences Image Denoising Using Spatial Aaptive Thresholing Raneesh Mishra M. Tech Stuent, Department of Electronics & Communication,

More information

Approximate Constraint Satisfaction Requires Large LP Relaxations

Approximate Constraint Satisfaction Requires Large LP Relaxations Approximate Constraint Satisfaction Requires Large LP Relaxations oah Fleming April 19, 2018 Linear programming is a very powerful tool for attacking optimization problems. Techniques such as the ellipsoi

More information

6 General properties of an autonomous system of two first order ODE

6 General properties of an autonomous system of two first order ODE 6 General properties of an autonomous system of two first orer ODE Here we embark on stuying the autonomous system of two first orer ifferential equations of the form ẋ 1 = f 1 (, x 2 ), ẋ 2 = f 2 (, x

More information

Leaving Randomness to Nature: d-dimensional Product Codes through the lens of Generalized-LDPC codes

Leaving Randomness to Nature: d-dimensional Product Codes through the lens of Generalized-LDPC codes Leaving Ranomness to Nature: -Dimensional Prouct Coes through the lens of Generalize-LDPC coes Tavor Baharav, Kannan Ramchanran Dept. of Electrical Engineering an Computer Sciences, U.C. Berkeley {tavorb,

More information

The Role of Models in Model-Assisted and Model- Dependent Estimation for Domains and Small Areas

The Role of Models in Model-Assisted and Model- Dependent Estimation for Domains and Small Areas The Role of Moels in Moel-Assiste an Moel- Depenent Estimation for Domains an Small Areas Risto Lehtonen University of Helsini Mio Myrsylä University of Pennsylvania Carl-Eri Särnal University of Montreal

More information

Perfect Matchings in Õ(n1.5 ) Time in Regular Bipartite Graphs

Perfect Matchings in Õ(n1.5 ) Time in Regular Bipartite Graphs Perfect Matchings in Õ(n1.5 ) Time in Regular Bipartite Graphs Ashish Goel Michael Kapralov Sanjeev Khanna Abstract We consier the well-stuie problem of fining a perfect matching in -regular bipartite

More information

Modeling of Dependence Structures in Risk Management and Solvency

Modeling of Dependence Structures in Risk Management and Solvency Moeling of Depenence Structures in Risk Management an Solvency University of California, Santa Barbara 0. August 007 Doreen Straßburger Structure. Risk Measurement uner Solvency II. Copulas 3. Depenent

More information

. Using a multinomial model gives us the following equation for P d. , with respect to same length term sequences.

. Using a multinomial model gives us the following equation for P d. , with respect to same length term sequences. S 63 Lecture 8 2/2/26 Lecturer Lillian Lee Scribes Peter Babinski, Davi Lin Basic Language Moeling Approach I. Special ase of LM-base Approach a. Recap of Formulas an Terms b. Fixing θ? c. About that Multinomial

More information

Binary Discrimination Methods for High Dimensional Data with a. Geometric Representation

Binary Discrimination Methods for High Dimensional Data with a. Geometric Representation Binary Discrimination Methos for High Dimensional Data with a Geometric Representation Ay Bolivar-Cime, Luis Miguel Corova-Roriguez Universia Juárez Autónoma e Tabasco, División Acaémica e Ciencias Básicas

More information

arxiv: v2 [cs.ds] 11 May 2016

arxiv: v2 [cs.ds] 11 May 2016 Optimizing Star-Convex Functions Jasper C.H. Lee Paul Valiant arxiv:5.04466v2 [cs.ds] May 206 Department of Computer Science Brown University {jasperchlee,paul_valiant}@brown.eu May 3, 206 Abstract We

More information

A simple tranformation of copulas

A simple tranformation of copulas A simple tranformation of copulas V. Durrleman, A. Nikeghbali & T. Roncalli Groupe e Recherche Opérationnelle Créit Lyonnais France July 31, 2000 Abstract We stuy how copulas properties are moifie after

More information

Linear Regression with Limited Observation

Linear Regression with Limited Observation Ela Hazan Tomer Koren Technion Israel Institute of Technology, Technion City 32000, Haifa, Israel ehazan@ie.technion.ac.il tomerk@cs.technion.ac.il Abstract We consier the most common variants of linear

More information

Practical Analysis of Key Recovery Attack against Search-LWE Problem

Practical Analysis of Key Recovery Attack against Search-LWE Problem Practical Analysis of Key Recovery Attack against Search-LWE Problem Royal Holloway an Kyushu University Workshop on Lattice-base cryptography 7 th September, 2016 Momonari Kuo Grauate School of Mathematics,

More information

Hybrid Fusion for Biometrics: Combining Score-level and Decision-level Fusion

Hybrid Fusion for Biometrics: Combining Score-level and Decision-level Fusion Hybri Fusion for Biometrics: Combining Score-level an Decision-level Fusion Qian Tao Raymon Velhuis Signals an Systems Group, University of Twente Postbus 217, 7500AE Enschee, the Netherlans {q.tao,r.n.j.velhuis}@ewi.utwente.nl

More information

Ramsey numbers of some bipartite graphs versus complete graphs

Ramsey numbers of some bipartite graphs versus complete graphs Ramsey numbers of some bipartite graphs versus complete graphs Tao Jiang, Michael Salerno Miami University, Oxfor, OH 45056, USA Abstract. The Ramsey number r(h, K n ) is the smallest positive integer

More information

New Statistical Test for Quality Control in High Dimension Data Set

New Statistical Test for Quality Control in High Dimension Data Set International Journal of Applie Engineering Research ISSN 973-456 Volume, Number 6 (7) pp. 64-649 New Statistical Test for Quality Control in High Dimension Data Set Shamshuritawati Sharif, Suzilah Ismail

More information

Math 1271 Solutions for Fall 2005 Final Exam

Math 1271 Solutions for Fall 2005 Final Exam Math 7 Solutions for Fall 5 Final Eam ) Since the equation + y = e y cannot be rearrange algebraically in orer to write y as an eplicit function of, we must instea ifferentiate this relation implicitly

More information