A Variational Approach to Semi-Supervised Clustering

Size: px
Start display at page:

Download "A Variational Approach to Semi-Supervised Clustering"

Transcription

1 A Variational Approach to Semi-Supervise Clustering Peng Li, Yiming Ying an Colin Campbell Department of Engineering Mathematics, University of Bristol, Bristol BS8 1TR, Unite Kingom Abstract We present a variational inference scheme for semi-supervise clustering in which ata is supplemente with sie information in the form of common labels. There is no mutual exclusion of classes assumption an samples are represente as a combinatorial mixture over multiple clusters. The metho has other avantages such as the ability to fin the most probable number of soft clusters in the ata, the ability to avoi overfitting an the ability to fin a confience measure for membership of a soft cluster. We illustrate performance of the metho on six ata sets an fin a positive comparison against constraine K-means clustering. 1 Introuction With semi-supervise clustering the aim is to fin clusters or meaningful partitions of the ata, aie by the use of sie information. This sie information can be in the form of must-links (two samples must be in the same cluster) an cannot-links (two samples must not belong to the same cluster). A number of approaches have been propose for semi-supervise clustering. Typically these methos have use the constraints from the sie information to either alter the objective function or the istance measure use. Many methos have use moifications of pre-existing clustering schemes such as incremental clustering algorithms 13], K-means clustering 2, 7] or Hien Markov Ranom Fiels 3]. One problem with some of these approaches is that the metho can work well if the correct number of clusters is alreay known: K-means clustering is an example. However, a principle approach to fining K is generally not given. Other potential isavantages of some clustering approaches is that there is an implicit mutual exclusion of clusters assumption i.e. samples are assume to uniquely belong to a particular cluster. In many cases this assumption may not be fully vali an it woul be more appropriate for a sample to be associate with multiple soft clusters. A further issue with some clustering approaches is that there is no principle mechanism to avoi fitting to noise in the ata. With Bayesian methos, however, we can incorporate a Bayes prior which penalizes such overfitting. Our motivation for consiering semi-supervise clustering comes from potential applications in cancer informatics. There have been a number of instances where unsupervise learning methos have been applie to cancer microarray ata sets, for example, an clinically istinct subtypes of cancer have been inicate e.g. 1, 9]. However, in some cases a specific causative event is known an thus it is possible to give common labels to some samples. A specific example might be acute lymphoblastic leukemia 15]. This isease is known to have a number of subtypes with variable response to chemotherapy. An originating event for some of these subtypes is believe known, sometimes stemming from the creation of fusion genes through genetic rearrangement involving genes BCR-ABL, E2A-PBX1, TEL-AML1 or rearrangements of the MLL gene. These rearrangements can be etecte via FISH (Fluorescent in situ Hybriization), an thus we can assign common labels to certain samples an use semi-supervise clustering to improve characterization of unlabelle samples. In these cancer applications, the sie information is typically in the form of must-links which will therefore be the focus of this paper. We thus propose a probabilistic graphical moel approach to semi-supervise clustering with the sie information given as the aition of must-links between some samples. The propose metho provies an objective measure of the number of clusters present an can reaily hanle missing values. Samples are represente as combinatorial mixtures over a set of latent processes or soft clusters, so there is no Current aress: Dept. of Computer Science, University College Lonon, Gower Street, Lonon WC1E 6BT, Unite Kingom 1

2 mutual exclusion of clusters assumption. In aition, by allowing a representation overlapping ifferent clusters we can erive a confience measure for cluster membership: a confience measure for subtype membership is important in clinical cancer informatics applications. As a Bayesian methoology we can also moify the metho to incorporate a Bayes prior penalizing over-complex moels which woul fit to noise in the ata. In the next two sections we propose our semi-supervise probabilistic graphical moel, followe by experiments in section 4. 2 The Methos Our semi-supervise moel utilizes the Latent Process Decomposition (LPD) moel evelope in 5, 12], an hence we will call this variant semi-supervise LPD or SLPD. For the natural numbers we aopt the notation N m = {1,..., m} for any m N. For the ata we use as the sample inex, g as the attribute inex, an script letters D, G to inex the corresponing number of samples an attributes. The number of clusters is K. The complete ata set is E = {E g : N D, g N G }. This notation stems from our cancer informatics motivation of using the expression value of gene g in sample. In our semi-supervise setting, we have aitional block information C where each block c enotes a set of ata points that is known to belong to a single class. In keeping with the stanar Bayesian moels, we also assume both blocks an the ata points in each block are i.i.. sample. Specifically, this sie information can be represente by a D C matrix δ with its entities δ c efine as follows { 1 if ata is a member of block c, δ c = 0 otherwise. We can now escribe our semi-supervise graphical moel. In probabilistic terms, the ata set E can be partitione into K-processes (soft clusters) escribe as follows. For a complete ata set, a Dirichlet prior istribution for the istribution of processes is efine by K-imensional parameter α. For each known block c, a istribution θ c over the set of mixture components inexe by k is rawn from a single Dirichlet istribution parameterize by α. Then, for all samples in block c (i.e. δ c = 1), the latent inicator variable Z g inicates which process k is chosen, with probability θ ck, from the common block-specific istribution θ c. The value E g for attribute g in sample is then rawn from the kth Gaussian with mean µ gk an eviation σ gk, enote as N(E g µ gk, σ gk ). We repeat the above proceure for each block in C. The graphical moel is illustrate in Figure 1 which is motivate by Latent Dirichlet Allocation 5]. Figure 1: A graphical moel representation of the metho propose in this paper. E g enotes the value of attribute g in sample. µ an σ are moel parameters. Z g is a hien variable giving the process inex of attribute g in sample. θ c gives the mixing over subgroups for sample in block c enote by c. The probability of θ c is given by a Dirichlet istribution with hyper-parameter α. The moel parameters are Θ = (µ, σ, α) an we use the notation c to enote sample in block c. From the graphical moel in Figure 1, we can formulate the block-specific joint istribution of the observe ata E an the latent variables Z by p(e, θ, Z Θ, C) = c p(θ c α) c p(e, Z θ c, Θ), (1) where p(θ c α) is Dirichlet efine by p(θ c α) = Γ( k α k) k Γ(α k) see that k θα k 1 ck. Using the block matrix δ, we further 2

3 p(e, Z θ c, Θ) = ] δc p(z θ c )p(e Z, µ, σ) c = δc g p(z g θ c )N(E g µ g, σ g, Z g )]. (2) For notational simplicity, we regar Z g as a unit-basis vector (Z g,1,..., Z g,k ) which transforms the process latent variable Z g = k to the unique vector Z g given by Z g,k = 1 an Z g,j = 0 for j k. Equivalently, the ranom variable Z g is istribute accoring to a multinomial probability efine by p(z g θ c ) = k θz g,k ck. Hence, the above equation can be rewritten as p(e, θ, Z Θ, C) = p(θ c α) Zg,k δ c θ ck N(E g µ gk, σ gk )]. (3) c g k With these priors, the final ata likelihoo can be obtaine by marginalizing out the latent variables θ an Z := {Z g : N D, g N G } p(e Θ, C) := p(e, θ, Z Θ, C)θ. (4) In particular, we can see from equations (2) an (3) that θ Z p(e Θ, C) = δc p(z g θ c )N(E g µ g, σ g, Z g )] p(θ c α)θ c c θ c Z g = Zg,k θ ck N(E g µ gk, σ gk )] p(θ c α)θ c c θ c Z g, c c,g,k = δc θ ck N(E g µ g, σ g, Z g )] p(θ c α)θ c. c θ c,g k We shoul mention that, without block information (i.e. δ = I D D ), the above equation is the exact likelihoo of Latent Process Decomposition given in 12]. 3 Moel inference an parameter learning We now consier moel inference an parameter estimation uner SLPD. The main inferential goal is to compute the posterior istribution of the hien variables p(θ, Z E, Θ, C). One irect metho is to use Bayes rule p(θ, Z E, Θ, C) = p(e,θ,z Θ,C) p(e Θ,C). This approach is usually intractable since this involves computationally intensive estimation of multi-integrals in the final likelihoo p(e θ, C). In this paper, we rely on variational inference methos 10, 11] which maximize a lower boun on the likelihoo p(e Θ, C) to estimate the moel parameters Θ an approximate p(θ, Z E, Θ, C) in a hypothesis family. One common hypothesis family is the factorize family efine by q(θ, Z γ, Q) = q(θ γ)q(z Q) with variational parameters γ, Q where, in the expression of the istribution q, the epenency on the E, Θ, C is omitte. More specifically, in our moel we assume that q(θ γ) = c q(θ c γ c ) = ( Γ( k γ ) ck) c k Γ(γ ck) k θγ ck 1 ck, an q(z Q) =,g q(z g Q g ) = ),g ( k QZ g,k g,k, among which γ, Q will be set as we escribe below. We can lower boun the log-likelihoo by applying Jensen s inequality to equation (4): log p(e Θ, C) = log p(e, θ, Z Θ, C)θ θ Z L(γ, Q; Θ) := E q log p(e, θ, Z Θ, C) ] Eq q(θ, Z γ, Q) ]. Consequently we can estimate the variational an moel parameters by alternative coorinate ascent methos known as a variational EM algorithm. The etails are summarize in the Appenix. (5) 3

4 E-step: maximize L with respect to the variational parameters γ, Q an get the upates Q g,k = N(E g µ gk, σ gk ) c exp(δ c(ψ(γ ck ) Ψ( k γ ck))) ] k N(E g µ gk, σ gk ) c exp(δ cψ(γ ck ) Ψ( k γ ], (6) ck)) an γ ck = α k +,g δ c Q g,k. (7) M-step: maximize L with respect to α, µ an σ an get that µ gk = Q g,ke g Q, σgk 2 = Q g,k(e g µ gk ) 2 g,k Q. (8) g,k an the parameter α is foun using a Newton-Raphson metho as given in the Appenix. The above iterative proceure is run until convergence (plateauing of the lower boun or the estimate likelihoo p(e Θ, C), see the Appenix for etails). Interpretation of the resultant moel is very similar to Latent Process Decomposition 12]. When normalize over k, the parameter γ ck gives the confience that a sample belonging to block c (which share a common label) belongs to soft cluster k. For each soft cluster k the moel parameters µ gk an σ gk give a ensity istribution for the attribute value g over all samples, see 6] for examples of use of these ensity estimators in application to interpreting breast cancer microarray ata. If some values of E g are missing, we omit corresponing contributions in the M-step upates an the corresponing Q g,k. The above argument is base on a maximum likelihoo approach. We en this section with a comment about the maximum a posterior (MAP) solution. In this case µ gk an σ gk are prouce by prior istributions p(µ) = gk p(µ gk) an p(σ 2 ) = gk p(σ2 gk ) with p(µ gk) N(0, σ µ ), p(σgk 2 ) exp( ) s σ. gk 2 These priors for penalizing over-complexity are justifie in 12]. In this case we only nee to change the upates for µ an σ µ gk = σ2 µ Q g,ke g σgk 2 + σ2 µ Q g,k), σ2 gk = Q g,k(e g µ gk ) 2 + 2s Q. (9) g,k The argument leaing to the above upates is the same as before except we replace the likelihoo p(e Θ) by p(e Θ)p(µ)p(σ), an consequently the lower boun is replace by L(Q; Θ) + log p(µ) + log p(σ). 4 Experimental Results To valiate the propose approach, we investigate the ML solution applie to four atasets from the UCI Repository 8] an two cancer microarray ata sets (see Table 1). The four ata sets from the UCI Repository have known sample labels an thus we can achieve an objective performance measure. As mentione in the introuction section, our interest in semi-supervise clustering stems from a potential use in cancer informatics. Thus the next two ata sets we consier are for cancer. One is a leukemia microarray ata 15]: in this case some labels are known since the causative events are observe genetic translocations or rearrangements. We have only use a subset of the original ata with unambiguous sample labels. The secon ataset is for lung cancer 4]. Again this consiste of histologically labelle samples erive from squamous cell lung carcinomas (21 samples), aenocarcinomas (139 samples), pulmonary carcinois (20 samples), small-cell lung carcinomas (6 samples) an normal lung tissue (17 samples). The imensionality of the both cancer atasets was reuce to 500 features base on largest variance. In the evaluation of the methos given below, we investigate three issues. Firstly, we pursue a comparison of the propose metho with pre-existing semi-supervise clustering methos. As our principal comparison we will use constraine K-means clustering (CKM) 2, 7] since this metho has been wiely use (we will only compare against constraine K-means clustering with must-links). Seconly, we will consier the possible gains to be mae by using sample label information. Finally, we also compare the unsupervise ULPD with semi-supervise SLPD to evaluate the gains mae by using sie information. We use the Balance Ran Inex (BRI) as evaluation criterion. Let nts (ntd) be the true number of pairs of ata in the same (ifferent) clusters an np S (np D) be the correctly preicte number of pairs of ata in the same (ifferent) clusters. BRI is efine as 4

5 Data Sets Letter {I, J} Wine Iris Digit {3, 6, 8} Leukemia Lung Cancer Number of Samples Number of Features Number of Classes Table 1: Data sets use in the experimental stuy. Data UKM CKM ULPD SLPD Sets 0 25% 50% 0 25% 50% Letter 0.501± ± ± ± ± ±0.039 Wine 0.877± ± ± ± ± ±0.032 Iris 0.824± ± ± ± ± ±0.038 Digit 0.751± ± ± ± ± ±0.045 Table 2: A comparison of constraine K-means clustering an SLPD. The entries are the BRI (mean ± stanar eviation over 100 trials using 3-fol cross valiation tests). Hypothesis testing inicates a statistically significant performance gain over CKM. ( nps BRI = 0.5 nts + npd ) ntd (10) The stanar Ran Inex is efine as (nps + npd)/(nts + ntd), which favors assigning ata points to ifferent clusters 14]. BRI favors matche pairs an mismatche pairs equally, hence it is more suitable for evaluating clustering algorithms. In Table 2 we tabulate performance for both constraine K-means clustering an SLPD using the BRI. For each stuy, the whole ata set was ivie into 3 fols by ranom partition. One fol of the ata set was use as a test set an a subset of the remaining two fols was use for supervise training. Exactly the same sample allocations were use in the evaluation of both constraine K-means clustering an SLPD. For the training set we also impose a egree of supervision to compare unsupervise learning with semi-supervise clustering using ifferent levels of supervision. The fraction of the ata set use for supervision was 0% (unsupervise), 25% an 50%. With ranom resampling of the ata, this 3-fol cross valiation proceure was repeate over 100 trials. As observe from Tables 2 an 3, SLPD compares favorably with CKM. In aition, by enforcing a egree of supervision (from 0% to 50%), we can observe performance improvements of both SLPD an CKM over LPD an unconstraine K-means clustering (UKM). As a real-life application we teste SLPD on two cancer microarray atasets for leukemia an lung cancer. We selecte these atasets since labels can be reliably assigne to a large proportion of the samples an thus we can evaluate performance of the propose metetho. Thus, for the leukemia ataset, we use a reuce 90-sample, 6-class subset of the ata where labels can be assigne in our stuy (balance to give 15 samples per class). However, for the full ataset there are a small number of samples with unclear assignment to subtype: this subset woul be a goo target for semi-supervise clustering since we woul want to use all available information to characterise these samples as efining novel subtypes or as having a relation to known subtypes. The results are tabulate in Table 3, where the values are the average values of 50 runs. For the leukemia ataset, both unsupervise LPD an semi-supervise LPD were traine on 45 samples from 3 classes, with clustering performance evaluate on the other 45 samples from the remaining 3 classes. For leukemia we fin that the BRI inex improves as we increase the extent of supervision. We also fin that SLPD consistently outperforms K-means clustering. A similar picture is repeate for lung cancer. Constraine K-means clustering is faster to execute compare to SLPD. However, SLPD has important avantages over constraine K-means. K-means clustering can work well if K is known but the metho oes not have an intrinsic mechanism for etermining the correct moel complexity. With SLPD we can etermine the appropriate number of clusters by etermining the estimate log-likelihoo on hol-out ata in a cross-valiation stuy (see Appenix 6.2). In Figure 2 we illustrate this proceure using the wine ata set mentione earlier. This ata set has 3 class labels an 178 samples. To etermine the appropriate moel complexity we fin the average log-likelihoo on 28 hol-out samples using the likelihoo estimate given in the Appenix (section 6.2, the curves are averages over 100 runs, error bars are omitte to simplify the Figure). For unsupervise learning the peak is at 4 (soli curve). This result appears reasonable since a enrogram ecomposition 5

6 Data UKM CKM ULPD SLPD Sets 0 25% 50% 0 25% 50% Leukemia 0.786± ± ± ± ± ±0.048 Lung Cancer 0.578± ± ± ± ± ±0.039 Table 3: A comparison of constraine K-means clustering an SLPD. The entries are the BRI (mean ± stanar eviation over the 50 trials of 3-fol cross valiation tests). Hypothesis testing inicates a statistically significant performance gain over CKM. of this ata set (Figure 2 (right)) suggests this is an appropriate ecomposition. If all the class labels are use we get a sharp peak at 3 (upper ashe curve) as expecte. However, if we use the labels of one class an leave the other two classes unlabelle we get a shallow peak at 3 (lower otte curve). This illustrates the point that the class labelling may not necessarily reflect the unerlying cluster structure, potentially negating any gains mae by using sie information. Log likelihoo Unsupervise LPD SLPD with 1 class labelle SLPD with all classes labelle x K number of clusters 0 Figure 2: The left-han Figure gives the Log-likelihoo (y-axis) versus K for the wine ata set for varying egrees of supervision (see text). The right-han Figure gives a enrogram partitioning of the same ata set using a correlation istance measure an average linkage (only the upper branches of the enrogram are illustrate). 5 Conclusion In this paper we have presente a novel variational inference metho for clustering with sie information. The metho provies a principle approach to moel selection via etermination of the log-likelihoo on hol-out ata, as illustrate in Figure 2. By contrast, many earlier methos for clustering with sie information appear to lack a soun approach to establishing the correct moel complexity. As a Bayesian methoology we coul also implement prior beliefs which can improve moel performance. Specifically, we can incorporate a Bayes prior penalizing overfitting to noise. Semi-supervise clustering methos have many important applications: we have mentione cancer applications where partial labelling of the samples is sometimes possible. Determining the number of subtypes an avoiing a fit to the extensive noise present in many of these ata sets are important issues in this application omain. 6 Appenix: erivation of the inference upates for SLPD 6.1 EM-upates In the E-step, we compute the posterior istribution q(e, θ γ, Q) an get the corresponing upates for γ, Q. To this en, we take the functional erivative of q, an get that q(z Q) exp ( E q(θ γ) log p(e, θ, Z Θ, C) ]), (11) an q(θ γ) exp ( E q(z Q) log p(e, θ, Z Θ, C) ]). (12) 6

7 Note that the log joint likelihoo can be expresse by log p(e, θ, Z Θ, C) = c,k (α k 1)log θ ck + C(log Γ( k α k) k log Γ(α k)) + c,,g,k Z g,kδ c log θ ck + c,,g,k Z g,kδ c log N(E g µ gk, σ gk ) = c,k (α k 1)log θ ck + C(log Γ( k α k) k log Γ(α k)) + c,,g,k Z g,kδ c log θ ck +,g,k Z g,k log N(E g µ gk, σ gk ), where we use the fact c δ ] c = 1 for any in the last equation. Note that E q(θc γ c) log θck = Ψ(γck ) Ψ( k γ ck) where Ψ is igamma function. Consequently, from the above log joint likelihoo equation we have that ] E q(θ γ) log p(e, θ, Z Θ, C) =,g,k g,k( Z c δ c(ψ(γ ck ) Ψ( ) k γ ck)) + log N(E g µ gk, σ gk, ] an E q(z Q) log p(e, θ, Z Θ, C) = c,k log θ ck( αk 1 +,g Q ) g,kδ c. Therefore, putting the above equations into equations (11), (12) an normalizing Q we have the esire upates (6) an (7) in Section 3. L At the M-step, the upates (8) for µ, σ are easy to get by solving the stationary equations µ gk = 0 an L σ 2 gk = 0. However, for the parameter α, observe that, in its erivative equation L α i = CΨ( k α k) CΨ(α i )+ c (Ψ(γ ci) Ψ( k γ ck)), the variable α i is mixe with other ones α j with j i, then we have to use iterative algorithm. Here, we employ a Newton-Raphson metho to approximate the optimal α. For this purpose, we compute the Hessian as follows H ij = C(Ψ ( k α k) δ ij Ψ (α i )). Hence, we have the iterative proceure α new = α ol ( H(α ol ) ) 1 L(α ol ). α Because of the special form of the Hessian matrix (ecomposable into iagonal an off-iagonal terms) we can avoi matrix inversion of H ij, see Appenix of 5]. 6.2 Computing the likelihoo Once the parameters have been estimate, we can calculate the likelihoo using equation (5). By the law of large numbers in probability theory, we can estimate the expectation with respect to p(θ c α) by averaging over large enough set of N samples: p(e Θ, C) = c 1 N ] δc N N(E g µ gk, σ gk )θ ckn. n=1 g k where {θ ckn } are N samples rawn from the Dirichlet istribution with parameter θ c. Hence References log p(e Θ, C) = c ] log 1 δc N(E g µ gk, σ gk )θ ckn N n = C log N + c g log n k ] δc N(E g µ gk, σ gk )θ ckn. 1] A.A. Alizaeh et al. Distinct types of iffuse large B-cell lymphoma ientifie by gene expression profiling. Nature, 403(3): , February ] S. Basu, A. Banerjee, an R. J. Mooney. Semi-supervise clustering by seeing. In Proceeings of International Conference on Machine Learning 2002, pages 27 34, g k 7

8 3] S. Basu, M. Bilenko, an R. J. Mooney. A probabilistic framework for semi-supervise clustering. In KDD 04: Proceeings of the tenth ACM SIGKDD international conference on Knowlege iscovery an ata mining, pages 59 68, New York, NY, USA, ACM Press. 4] A Bhattacharjee et al. Classification of human lung carcinomas by mrna expression profiling reveals istinct aenocarcinoma subclasses. Proceeings of the National Acaemy of Sciences, 98: , ] D. M. Blei, Anrew Y.Ng, an M. I. Joran. Latent irichlet allocation. Journal of Machine Learning Research, 3: , ] L. Carrivick, S. Rogers, J. Clark, C. Campbell, M. Girolami, an C. Cooper. Ientification of prognostic signatures in breast cancer microarray ata using bayesian techniques. Journal of the Royal Society: Interface, 3: , ] A. Demiriz, K. Bennett, an M. Embrechts. Semi-supervise clustering using genetic algorithms, ] C.L. Blake D.J. Newman, S. Hettich an C.J. Merz. UCI repository of machine learning atabases, mlearn/mlrepository.html, ] E. Garber et al. Diversity of gene expression in aenocarcinoma of the lung. Proceeings National Acaemy Sciences, 98(24): , ] Z. Ghahramani an M. J. Beal. Graphical moels an variational methos. MIT Press, ] M. I. Joran, Z. Ghahramani, T. Jaakkola, an L. K. Saul. An introuction to variational methos for graphical moels. Machine Learning, 37: , ] S. Rogers, M. Girolami, C. Campbell, an R. Breitling. The latent process ecomposition of cdna microarray atasets. IEEE/ACM Transactions on Computational Biology an Bioinformatics, 2: , ] K. Wagstaff an C. Carie. Clustering with instance-level constraints. In Proceeings of the Seventeenth International Conference on Machine Learning, pages , ] E. Xing, A. Ng, M. Joran, an S. Russell. Distance metric learning, with application to clustering with sie-information. In Avances in NIPS, number vol. 15, ] E-J Yeoh et al. Classification, subtype iscovery, an preiction of outcome in peiatric acute lymphoblastic leukemia by gene expression profiling. Cancer Cell, 1: ,

Survey Sampling. 1 Design-based Inference. Kosuke Imai Department of Politics, Princeton University. February 19, 2013

Survey Sampling. 1 Design-based Inference. Kosuke Imai Department of Politics, Princeton University. February 19, 2013 Survey Sampling Kosuke Imai Department of Politics, Princeton University February 19, 2013 Survey sampling is one of the most commonly use ata collection methos for social scientists. We begin by escribing

More information

LDA Collapsed Gibbs Sampler, VariaNonal Inference. Task 3: Mixed Membership Models. Case Study 5: Mixed Membership Modeling

LDA Collapsed Gibbs Sampler, VariaNonal Inference. Task 3: Mixed Membership Models. Case Study 5: Mixed Membership Modeling Case Stuy 5: Mixe Membership Moeling LDA Collapse Gibbs Sampler, VariaNonal Inference Machine Learning for Big Data CSE547/STAT548, University of Washington Emily Fox May 8 th, 05 Emily Fox 05 Task : Mixe

More information

This module is part of the. Memobust Handbook. on Methodology of Modern Business Statistics

This module is part of the. Memobust Handbook. on Methodology of Modern Business Statistics This moule is part of the Memobust Hanbook on Methoology of Moern Business Statistics 26 March 2014 Metho: Balance Sampling for Multi-Way Stratification Contents General section... 3 1. Summary... 3 2.

More information

Cascaded redundancy reduction

Cascaded redundancy reduction Network: Comput. Neural Syst. 9 (1998) 73 84. Printe in the UK PII: S0954-898X(98)88342-5 Cascae reunancy reuction Virginia R e Sa an Geoffrey E Hinton Department of Computer Science, University of Toronto,

More information

Computing Exact Confidence Coefficients of Simultaneous Confidence Intervals for Multinomial Proportions and their Functions

Computing Exact Confidence Coefficients of Simultaneous Confidence Intervals for Multinomial Proportions and their Functions Working Paper 2013:5 Department of Statistics Computing Exact Confience Coefficients of Simultaneous Confience Intervals for Multinomial Proportions an their Functions Shaobo Jin Working Paper 2013:5

More information

Time-of-Arrival Estimation in Non-Line-Of-Sight Environments

Time-of-Arrival Estimation in Non-Line-Of-Sight Environments 2 Conference on Information Sciences an Systems, The Johns Hopkins University, March 2, 2 Time-of-Arrival Estimation in Non-Line-Of-Sight Environments Sinan Gezici, Hisashi Kobayashi an H. Vincent Poor

More information

Collapsed Gibbs and Variational Methods for LDA. Example Collapsed MoG Sampling

Collapsed Gibbs and Variational Methods for LDA. Example Collapsed MoG Sampling Case Stuy : Document Retrieval Collapse Gibbs an Variational Methos for LDA Machine Learning/Statistics for Big Data CSE599C/STAT59, University of Washington Emily Fox 0 Emily Fox February 7 th, 0 Example

More information

Parameter estimation: A new approach to weighting a priori information

Parameter estimation: A new approach to weighting a priori information Parameter estimation: A new approach to weighting a priori information J.L. Mea Department of Mathematics, Boise State University, Boise, ID 83725-555 E-mail: jmea@boisestate.eu Abstract. We propose a

More information

Tutorial on Maximum Likelyhood Estimation: Parametric Density Estimation

Tutorial on Maximum Likelyhood Estimation: Parametric Density Estimation Tutorial on Maximum Likelyhoo Estimation: Parametric Density Estimation Suhir B Kylasa 03/13/2014 1 Motivation Suppose one wishes to etermine just how biase an unfair coin is. Call the probability of tossing

More information

Topic Modeling: Beyond Bag-of-Words

Topic Modeling: Beyond Bag-of-Words Hanna M. Wallach Cavenish Laboratory, University of Cambrige, Cambrige CB3 0HE, UK hmw26@cam.ac.u Abstract Some moels of textual corpora employ text generation methos involving n-gram statistics, while

More information

A Review of Multiple Try MCMC algorithms for Signal Processing

A Review of Multiple Try MCMC algorithms for Signal Processing A Review of Multiple Try MCMC algorithms for Signal Processing Luca Martino Image Processing Lab., Universitat e València (Spain) Universia Carlos III e Mari, Leganes (Spain) Abstract Many applications

More information

'HVLJQ &RQVLGHUDWLRQ LQ 0DWHULDO 6HOHFWLRQ 'HVLJQ 6HQVLWLYLW\,1752'8&7,21

'HVLJQ &RQVLGHUDWLRQ LQ 0DWHULDO 6HOHFWLRQ 'HVLJQ 6HQVLWLYLW\,1752'8&7,21 Large amping in a structural material may be either esirable or unesirable, epening on the engineering application at han. For example, amping is a esirable property to the esigner concerne with limiting

More information

Collapsed Variational Inference for HDP

Collapsed Variational Inference for HDP Collapse Variational Inference for HDP Yee W. Teh Davi Newman an Max Welling Publishe on NIPS 2007 Discussion le by Iulian Pruteanu Outline Introuction Hierarchical Bayesian moel for LDA Collapse VB inference

More information

Robust Forward Algorithms via PAC-Bayes and Laplace Distributions. ω Q. Pr (y(ω x) < 0) = Pr A k

Robust Forward Algorithms via PAC-Bayes and Laplace Distributions. ω Q. Pr (y(ω x) < 0) = Pr A k A Proof of Lemma 2 B Proof of Lemma 3 Proof: Since the support of LL istributions is R, two such istributions are equivalent absolutely continuous with respect to each other an the ivergence is well-efine

More information

A Modification of the Jarque-Bera Test. for Normality

A Modification of the Jarque-Bera Test. for Normality Int. J. Contemp. Math. Sciences, Vol. 8, 01, no. 17, 84-85 HIKARI Lt, www.m-hikari.com http://x.oi.org/10.1988/ijcms.01.9106 A Moification of the Jarque-Bera Test for Normality Moawa El-Fallah Ab El-Salam

More information

Robust Low Rank Kernel Embeddings of Multivariate Distributions

Robust Low Rank Kernel Embeddings of Multivariate Distributions Robust Low Rank Kernel Embeings of Multivariate Distributions Le Song, Bo Dai College of Computing, Georgia Institute of Technology lsong@cc.gatech.eu, boai@gatech.eu Abstract Kernel embeing of istributions

More information

Schrödinger s equation.

Schrödinger s equation. Physics 342 Lecture 5 Schröinger s Equation Lecture 5 Physics 342 Quantum Mechanics I Wenesay, February 3r, 2010 Toay we iscuss Schröinger s equation an show that it supports the basic interpretation of

More information

Least-Squares Regression on Sparse Spaces

Least-Squares Regression on Sparse Spaces Least-Squares Regression on Sparse Spaces Yuri Grinberg, Mahi Milani Far, Joelle Pineau School of Computer Science McGill University Montreal, Canaa {ygrinb,mmilan1,jpineau}@cs.mcgill.ca 1 Introuction

More information

Lecture 2: Correlated Topic Model

Lecture 2: Correlated Topic Model Probabilistic Moels for Unsupervise Learning Spring 203 Lecture 2: Correlate Topic Moel Inference for Correlate Topic Moel Yuan Yuan First of all, let us make some claims about the parameters an variables

More information

Research Article When Inflation Causes No Increase in Claim Amounts

Research Article When Inflation Causes No Increase in Claim Amounts Probability an Statistics Volume 2009, Article ID 943926, 10 pages oi:10.1155/2009/943926 Research Article When Inflation Causes No Increase in Claim Amounts Vytaras Brazauskas, 1 Bruce L. Jones, 2 an

More information

Lower Bounds for the Smoothed Number of Pareto optimal Solutions

Lower Bounds for the Smoothed Number of Pareto optimal Solutions Lower Bouns for the Smoothe Number of Pareto optimal Solutions Tobias Brunsch an Heiko Röglin Department of Computer Science, University of Bonn, Germany brunsch@cs.uni-bonn.e, heiko@roeglin.org Abstract.

More information

Introduction to Machine Learning

Introduction to Machine Learning How o you estimate p(y x)? Outline Contents Introuction to Machine Learning Logistic Regression Varun Chanola April 9, 207 Generative vs. Discriminative Classifiers 2 Logistic Regression 2 3 Logistic Regression

More information

Latent Dirichlet Allocation in Web Spam Filtering

Latent Dirichlet Allocation in Web Spam Filtering Latent Dirichlet Allocation in Web Spam Filtering István Bíró Jácint Szabó Anrás A. Benczúr Data Mining an Web search Research Group, Informatics Laboratory Computer an Automation Research Institute of

More information

New Statistical Test for Quality Control in High Dimension Data Set

New Statistical Test for Quality Control in High Dimension Data Set International Journal of Applie Engineering Research ISSN 973-456 Volume, Number 6 (7) pp. 64-649 New Statistical Test for Quality Control in High Dimension Data Set Shamshuritawati Sharif, Suzilah Ismail

More information

Model-based clustering and validation techniques for gene expression data

Model-based clustering and validation techniques for gene expression data Moel-base clustering an valiation techniques for gene expression ata Ka Yee Yeung Department of Microbiology University of Washington, Seattle WA Overview Introuction to microarrays Clustering 11 Valiating

More information

Optimization of Geometries by Energy Minimization

Optimization of Geometries by Energy Minimization Optimization of Geometries by Energy Minimization by Tracy P. Hamilton Department of Chemistry University of Alabama at Birmingham Birmingham, AL 3594-140 hamilton@uab.eu Copyright Tracy P. Hamilton, 1997.

More information

Hybrid Fusion for Biometrics: Combining Score-level and Decision-level Fusion

Hybrid Fusion for Biometrics: Combining Score-level and Decision-level Fusion Hybri Fusion for Biometrics: Combining Score-level an Decision-level Fusion Qian Tao Raymon Velhuis Signals an Systems Group, University of Twente Postbus 217, 7500AE Enschee, the Netherlans {q.tao,r.n.j.velhuis}@ewi.utwente.nl

More information

Influence of weight initialization on multilayer perceptron performance

Influence of weight initialization on multilayer perceptron performance Influence of weight initialization on multilayer perceptron performance M. Karouia (1,2) T. Denœux (1) R. Lengellé (1) (1) Université e Compiègne U.R.A. CNRS 817 Heuiasyc BP 649 - F-66 Compiègne ceex -

More information

Proof of SPNs as Mixture of Trees

Proof of SPNs as Mixture of Trees A Proof of SPNs as Mixture of Trees Theorem 1. If T is an inuce SPN from a complete an ecomposable SPN S, then T is a tree that is complete an ecomposable. Proof. Argue by contraiction that T is not a

More information

Estimating Causal Direction and Confounding Of Two Discrete Variables

Estimating Causal Direction and Confounding Of Two Discrete Variables Estimating Causal Direction an Confouning Of Two Discrete Variables This inspire further work on the so calle aitive noise moels. Hoyer et al. (2009) extene Shimizu s ientifiaarxiv:1611.01504v1 [stat.ml]

More information

Hyperbolic Moment Equations Using Quadrature-Based Projection Methods

Hyperbolic Moment Equations Using Quadrature-Based Projection Methods Hyperbolic Moment Equations Using Quarature-Base Projection Methos J. Koellermeier an M. Torrilhon Department of Mathematics, RWTH Aachen University, Aachen, Germany Abstract. Kinetic equations like the

More information

On the Surprising Behavior of Distance Metrics in High Dimensional Space

On the Surprising Behavior of Distance Metrics in High Dimensional Space On the Surprising Behavior of Distance Metrics in High Dimensional Space Charu C. Aggarwal, Alexaner Hinneburg 2, an Daniel A. Keim 2 IBM T. J. Watson Research Center Yortown Heights, NY 0598, USA. charu@watson.ibm.com

More information

Lagrangian and Hamiltonian Mechanics

Lagrangian and Hamiltonian Mechanics Lagrangian an Hamiltonian Mechanics.G. Simpson, Ph.. epartment of Physical Sciences an Engineering Prince George s Community College ecember 5, 007 Introuction In this course we have been stuying classical

More information

Similarity Measures for Categorical Data A Comparative Study. Technical Report

Similarity Measures for Categorical Data A Comparative Study. Technical Report Similarity Measures for Categorical Data A Comparative Stuy Technical Report Department of Computer Science an Engineering University of Minnesota 4-92 EECS Builing 200 Union Street SE Minneapolis, MN

More information

Multi-View Clustering via Canonical Correlation Analysis

Multi-View Clustering via Canonical Correlation Analysis Technical Report TTI-TR-2008-5 Multi-View Clustering via Canonical Correlation Analysis Kamalika Chauhuri UC San Diego Sham M. Kakae Toyota Technological Institute at Chicago ABSTRACT Clustering ata in

More information

The total derivative. Chapter Lagrangian and Eulerian approaches

The total derivative. Chapter Lagrangian and Eulerian approaches Chapter 5 The total erivative 51 Lagrangian an Eulerian approaches The representation of a flui through scalar or vector fiels means that each physical quantity uner consieration is escribe as a function

More information

Generalizing Kronecker Graphs in order to Model Searchable Networks

Generalizing Kronecker Graphs in order to Model Searchable Networks Generalizing Kronecker Graphs in orer to Moel Searchable Networks Elizabeth Boine, Babak Hassibi, Aam Wierman California Institute of Technology Pasaena, CA 925 Email: {eaboine, hassibi, aamw}@caltecheu

More information

Thermal conductivity of graded composites: Numerical simulations and an effective medium approximation

Thermal conductivity of graded composites: Numerical simulations and an effective medium approximation JOURNAL OF MATERIALS SCIENCE 34 (999)5497 5503 Thermal conuctivity of grae composites: Numerical simulations an an effective meium approximation P. M. HUI Department of Physics, The Chinese University

More information

Multi-View Clustering via Canonical Correlation Analysis

Multi-View Clustering via Canonical Correlation Analysis Keywors: multi-view learning, clustering, canonical correlation analysis Abstract Clustering ata in high-imensions is believe to be a har problem in general. A number of efficient clustering algorithms

More information

Optimal Signal Detection for False Track Discrimination

Optimal Signal Detection for False Track Discrimination Optimal Signal Detection for False Track Discrimination Thomas Hanselmann Darko Mušicki Dept. of Electrical an Electronic Eng. Dept. of Electrical an Electronic Eng. The University of Melbourne The University

More information

The Press-Schechter mass function

The Press-Schechter mass function The Press-Schechter mass function To state the obvious: It is important to relate our theories to what we can observe. We have looke at linear perturbation theory, an we have consiere a simple moel for

More information

Topic 7: Convergence of Random Variables

Topic 7: Convergence of Random Variables Topic 7: Convergence of Ranom Variables Course 003, 2016 Page 0 The Inference Problem So far, our starting point has been a given probability space (S, F, P). We now look at how to generate information

More information

Lecture Introduction. 2 Examples of Measure Concentration. 3 The Johnson-Lindenstrauss Lemma. CS-621 Theory Gems November 28, 2012

Lecture Introduction. 2 Examples of Measure Concentration. 3 The Johnson-Lindenstrauss Lemma. CS-621 Theory Gems November 28, 2012 CS-6 Theory Gems November 8, 0 Lecture Lecturer: Alesaner Mąry Scribes: Alhussein Fawzi, Dorina Thanou Introuction Toay, we will briefly iscuss an important technique in probability theory measure concentration

More information

Concentration of Measure Inequalities for Compressive Toeplitz Matrices with Applications to Detection and System Identification

Concentration of Measure Inequalities for Compressive Toeplitz Matrices with Applications to Detection and System Identification Concentration of Measure Inequalities for Compressive Toeplitz Matrices with Applications to Detection an System Ientification Borhan M Sananaji, Tyrone L Vincent, an Michael B Wakin Abstract In this paper,

More information

Lecture 2 Lagrangian formulation of classical mechanics Mechanics

Lecture 2 Lagrangian formulation of classical mechanics Mechanics Lecture Lagrangian formulation of classical mechanics 70.00 Mechanics Principle of stationary action MATH-GA To specify a motion uniquely in classical mechanics, it suffices to give, at some time t 0,

More information

. Using a multinomial model gives us the following equation for P d. , with respect to same length term sequences.

. Using a multinomial model gives us the following equation for P d. , with respect to same length term sequences. S 63 Lecture 8 2/2/26 Lecturer Lillian Lee Scribes Peter Babinski, Davi Lin Basic Language Moeling Approach I. Special ase of LM-base Approach a. Recap of Formulas an Terms b. Fixing θ? c. About that Multinomial

More information

CONTROL CHARTS FOR VARIABLES

CONTROL CHARTS FOR VARIABLES UNIT CONTOL CHATS FO VAIABLES Structure.1 Introuction Objectives. Control Chart Technique.3 Control Charts for Variables.4 Control Chart for Mean(-Chart).5 ange Chart (-Chart).6 Stanar Deviation Chart

More information

Non-Linear Bayesian CBRN Source Term Estimation

Non-Linear Bayesian CBRN Source Term Estimation Non-Linear Bayesian CBRN Source Term Estimation Peter Robins Hazar Assessment, Simulation an Preiction Group Dstl Porton Down, UK. probins@stl.gov.uk Paul Thomas Hazar Assessment, Simulation an Preiction

More information

Convergence of Random Walks

Convergence of Random Walks Chapter 16 Convergence of Ranom Walks This lecture examines the convergence of ranom walks to the Wiener process. This is very important both physically an statistically, an illustrates the utility of

More information

u!i = a T u = 0. Then S satisfies

u!i = a T u = 0. Then S satisfies Deterministic Conitions for Subspace Ientifiability from Incomplete Sampling Daniel L Pimentel-Alarcón, Nigel Boston, Robert D Nowak University of Wisconsin-Maison Abstract Consier an r-imensional subspace

More information

A New Minimum Description Length

A New Minimum Description Length A New Minimum Description Length Soosan Beheshti, Munther A. Dahleh Laboratory for Information an Decision Systems Massachusetts Institute of Technology soosan@mit.eu,ahleh@lis.mit.eu Abstract The minimum

More information

Level Construction of Decision Trees in a Partition-based Framework for Classification

Level Construction of Decision Trees in a Partition-based Framework for Classification Level Construction of Decision Trees in a Partition-base Framework for Classification Y.Y. Yao, Y. Zhao an J.T. Yao Department of Computer Science, University of Regina Regina, Saskatchewan, Canaa S4S

More information

LATTICE-BASED D-OPTIMUM DESIGN FOR FOURIER REGRESSION

LATTICE-BASED D-OPTIMUM DESIGN FOR FOURIER REGRESSION The Annals of Statistics 1997, Vol. 25, No. 6, 2313 2327 LATTICE-BASED D-OPTIMUM DESIGN FOR FOURIER REGRESSION By Eva Riccomagno, 1 Rainer Schwabe 2 an Henry P. Wynn 1 University of Warwick, Technische

More information

Flexible High-Dimensional Classification Machines and Their Asymptotic Properties

Flexible High-Dimensional Classification Machines and Their Asymptotic Properties Journal of Machine Learning Research 16 (2015) 1547-1572 Submitte 1/14; Revise 9/14; Publishe 8/15 Flexible High-Dimensional Classification Machines an Their Asymptotic Properties Xingye Qiao Department

More information

Sparse Reconstruction of Systems of Ordinary Differential Equations

Sparse Reconstruction of Systems of Ordinary Differential Equations Sparse Reconstruction of Systems of Orinary Differential Equations Manuel Mai a, Mark D. Shattuck b,c, Corey S. O Hern c,a,,e, a Department of Physics, Yale University, New Haven, Connecticut 06520, USA

More information

The Exact Form and General Integrating Factors

The Exact Form and General Integrating Factors 7 The Exact Form an General Integrating Factors In the previous chapters, we ve seen how separable an linear ifferential equations can be solve using methos for converting them to forms that can be easily

More information

A Hybrid Approach for Modeling High Dimensional Medical Data

A Hybrid Approach for Modeling High Dimensional Medical Data A Hybri Approach for Moeling High Dimensional Meical Data Alok Sharma 1, Gofrey C. Onwubolu 1 1 University of the South Pacific, Fii sharma_al@usp.ac.f, onwubolu_g@usp.ac.f Abstract. his work presents

More information

7.1 Support Vector Machine

7.1 Support Vector Machine 67577 Intro. to Machine Learning Fall semester, 006/7 Lecture 7: Support Vector Machines an Kernel Functions II Lecturer: Amnon Shashua Scribe: Amnon Shashua 7. Support Vector Machine We return now to

More information

Separation of Variables

Separation of Variables Physics 342 Lecture 1 Separation of Variables Lecture 1 Physics 342 Quantum Mechanics I Monay, January 25th, 2010 There are three basic mathematical tools we nee, an then we can begin working on the physical

More information

Homework 2 EM, Mixture Models, PCA, Dualitys

Homework 2 EM, Mixture Models, PCA, Dualitys Homework 2 EM, Mixture Moels, PCA, Dualitys CMU 10-715: Machine Learning (Fall 2015) http://www.cs.cmu.eu/~bapoczos/classes/ml10715_2015fall/ OUT: Oct 5, 2015 DUE: Oct 19, 2015, 10:20 AM Guielines The

More information

Estimation of the Maximum Domination Value in Multi-Dimensional Data Sets

Estimation of the Maximum Domination Value in Multi-Dimensional Data Sets Proceeings of the 4th East-European Conference on Avances in Databases an Information Systems ADBIS) 200 Estimation of the Maximum Domination Value in Multi-Dimensional Data Sets Eleftherios Tiakas, Apostolos.

More information

Math Notes on differentials, the Chain Rule, gradients, directional derivative, and normal vectors

Math Notes on differentials, the Chain Rule, gradients, directional derivative, and normal vectors Math 18.02 Notes on ifferentials, the Chain Rule, graients, irectional erivative, an normal vectors Tangent plane an linear approximation We efine the partial erivatives of f( xy, ) as follows: f f( x+

More information

Analyzing Tensor Power Method Dynamics in Overcomplete Regime

Analyzing Tensor Power Method Dynamics in Overcomplete Regime Journal of Machine Learning Research 18 (2017) 1-40 Submitte 9/15; Revise 11/16; Publishe 4/17 Analyzing Tensor Power Metho Dynamics in Overcomplete Regime Animashree Ananumar Department of Electrical

More information

Structural Risk Minimization over Data-Dependent Hierarchies

Structural Risk Minimization over Data-Dependent Hierarchies Structural Risk Minimization over Data-Depenent Hierarchies John Shawe-Taylor Department of Computer Science Royal Holloway an Befor New College University of Lonon Egham, TW20 0EX, UK jst@cs.rhbnc.ac.uk

More information

Multi-View Clustering via Canonical Correlation Analysis

Multi-View Clustering via Canonical Correlation Analysis Kamalika Chauhuri ITA, UC San Diego, 9500 Gilman Drive, La Jolla, CA Sham M. Kakae Karen Livescu Karthik Sriharan Toyota Technological Institute at Chicago, 6045 S. Kenwoo Ave., Chicago, IL kamalika@soe.ucs.eu

More information

arxiv:hep-th/ v1 3 Feb 1993

arxiv:hep-th/ v1 3 Feb 1993 NBI-HE-9-89 PAR LPTHE 9-49 FTUAM 9-44 November 99 Matrix moel calculations beyon the spherical limit arxiv:hep-th/93004v 3 Feb 993 J. Ambjørn The Niels Bohr Institute Blegamsvej 7, DK-00 Copenhagen Ø,

More information

Chapter 6: Energy-Momentum Tensors

Chapter 6: Energy-Momentum Tensors 49 Chapter 6: Energy-Momentum Tensors This chapter outlines the general theory of energy an momentum conservation in terms of energy-momentum tensors, then applies these ieas to the case of Bohm's moel.

More information

Acute sets in Euclidean spaces

Acute sets in Euclidean spaces Acute sets in Eucliean spaces Viktor Harangi April, 011 Abstract A finite set H in R is calle an acute set if any angle etermine by three points of H is acute. We examine the maximal carinality α() of

More information

Situation awareness of power system based on static voltage security region

Situation awareness of power system based on static voltage security region The 6th International Conference on Renewable Power Generation (RPG) 19 20 October 2017 Situation awareness of power system base on static voltage security region Fei Xiao, Zi-Qing Jiang, Qian Ai, Ran

More information

On the Aloha throughput-fairness tradeoff

On the Aloha throughput-fairness tradeoff On the Aloha throughput-fairness traeoff 1 Nan Xie, Member, IEEE, an Steven Weber, Senior Member, IEEE Abstract arxiv:1605.01557v1 [cs.it] 5 May 2016 A well-known inner boun of the stability region of

More information

A PAC-Bayesian Approach to Spectrally-Normalized Margin Bounds for Neural Networks

A PAC-Bayesian Approach to Spectrally-Normalized Margin Bounds for Neural Networks A PAC-Bayesian Approach to Spectrally-Normalize Margin Bouns for Neural Networks Behnam Neyshabur, Srinah Bhojanapalli, Davi McAllester, Nathan Srebro Toyota Technological Institute at Chicago {bneyshabur,

More information

ALGEBRAIC AND ANALYTIC PROPERTIES OF ARITHMETIC FUNCTIONS

ALGEBRAIC AND ANALYTIC PROPERTIES OF ARITHMETIC FUNCTIONS ALGEBRAIC AND ANALYTIC PROPERTIES OF ARITHMETIC FUNCTIONS MARK SCHACHNER Abstract. When consiere as an algebraic space, the set of arithmetic functions equippe with the operations of pointwise aition an

More information

THE EFFICIENCIES OF THE SPATIAL MEDIAN AND SPATIAL SIGN COVARIANCE MATRIX FOR ELLIPTICALLY SYMMETRIC DISTRIBUTIONS

THE EFFICIENCIES OF THE SPATIAL MEDIAN AND SPATIAL SIGN COVARIANCE MATRIX FOR ELLIPTICALLY SYMMETRIC DISTRIBUTIONS THE EFFICIENCIES OF THE SPATIAL MEDIAN AND SPATIAL SIGN COVARIANCE MATRIX FOR ELLIPTICALLY SYMMETRIC DISTRIBUTIONS BY ANDREW F. MAGYAR A issertation submitte to the Grauate School New Brunswick Rutgers,

More information

Simulation of Angle Beam Ultrasonic Testing with a Personal Computer

Simulation of Angle Beam Ultrasonic Testing with a Personal Computer Key Engineering Materials Online: 4-8-5 I: 66-9795, Vols. 7-73, pp 38-33 oi:.48/www.scientific.net/kem.7-73.38 4 rans ech ublications, witzerlan Citation & Copyright (to be inserte by the publisher imulation

More information

Image Denoising Using Spatial Adaptive Thresholding

Image Denoising Using Spatial Adaptive Thresholding International Journal of Engineering Technology, Management an Applie Sciences Image Denoising Using Spatial Aaptive Thresholing Raneesh Mishra M. Tech Stuent, Department of Electronics & Communication,

More information

NOTES ON EULER-BOOLE SUMMATION (1) f (l 1) (n) f (l 1) (m) + ( 1)k 1 k! B k (y) f (k) (y) dy,

NOTES ON EULER-BOOLE SUMMATION (1) f (l 1) (n) f (l 1) (m) + ( 1)k 1 k! B k (y) f (k) (y) dy, NOTES ON EULER-BOOLE SUMMATION JONATHAN M BORWEIN, NEIL J CALKIN, AND DANTE MANNA Abstract We stuy a connection between Euler-MacLaurin Summation an Boole Summation suggeste in an AMM note from 196, which

More information

arxiv: v4 [math.pr] 27 Jul 2016

arxiv: v4 [math.pr] 27 Jul 2016 The Asymptotic Distribution of the Determinant of a Ranom Correlation Matrix arxiv:309768v4 mathpr] 7 Jul 06 AM Hanea a, & GF Nane b a Centre of xcellence for Biosecurity Risk Analysis, University of Melbourne,

More information

05 The Continuum Limit and the Wave Equation

05 The Continuum Limit and the Wave Equation Utah State University DigitalCommons@USU Founations of Wave Phenomena Physics, Department of 1-1-2004 05 The Continuum Limit an the Wave Equation Charles G. Torre Department of Physics, Utah State University,

More information

Lower bounds on Locality Sensitive Hashing

Lower bounds on Locality Sensitive Hashing Lower bouns on Locality Sensitive Hashing Rajeev Motwani Assaf Naor Rina Panigrahy Abstract Given a metric space (X, X ), c 1, r > 0, an p, q [0, 1], a istribution over mappings H : X N is calle a (r,

More information

Estimation of District Level Poor Households in the State of. Uttar Pradesh in India by Combining NSSO Survey and

Estimation of District Level Poor Households in the State of. Uttar Pradesh in India by Combining NSSO Survey and Int. Statistical Inst.: Proc. 58th Worl Statistical Congress, 2011, Dublin (Session CPS039) p.6567 Estimation of District Level Poor Househols in the State of Uttar Praesh in Inia by Combining NSSO Survey

More information

FLUCTUATIONS IN THE NUMBER OF POINTS ON SMOOTH PLANE CURVES OVER FINITE FIELDS. 1. Introduction

FLUCTUATIONS IN THE NUMBER OF POINTS ON SMOOTH PLANE CURVES OVER FINITE FIELDS. 1. Introduction FLUCTUATIONS IN THE NUMBER OF POINTS ON SMOOTH PLANE CURVES OVER FINITE FIELDS ALINA BUCUR, CHANTAL DAVID, BROOKE FEIGON, MATILDE LALÍN 1 Introuction In this note, we stuy the fluctuations in the number

More information

Homework 2 Solutions EM, Mixture Models, PCA, Dualitys

Homework 2 Solutions EM, Mixture Models, PCA, Dualitys Homewor Solutions EM, Mixture Moels, PCA, Dualitys CMU 0-75: Machine Learning Fall 05 http://www.cs.cmu.eu/~bapoczos/classes/ml075_05fall/ OUT: Oct 5, 05 DUE: Oct 9, 05, 0:0 AM An EM algorithm for a Mixture

More information

Distributed coordination control for multi-robot networks using Lyapunov-like barrier functions

Distributed coordination control for multi-robot networks using Lyapunov-like barrier functions IEEE TRANSACTIONS ON 1 Distribute coorination control for multi-robot networks using Lyapunov-like barrier functions Dimitra Panagou, Dušan M. Stipanović an Petros G. Voulgaris Abstract This paper aresses

More information

Linear First-Order Equations

Linear First-Order Equations 5 Linear First-Orer Equations Linear first-orer ifferential equations make up another important class of ifferential equations that commonly arise in applications an are relatively easy to solve (in theory)

More information

TEMPORAL AND TIME-FREQUENCY CORRELATION-BASED BLIND SOURCE SEPARATION METHODS. Yannick DEVILLE

TEMPORAL AND TIME-FREQUENCY CORRELATION-BASED BLIND SOURCE SEPARATION METHODS. Yannick DEVILLE TEMPORAL AND TIME-FREQUENCY CORRELATION-BASED BLIND SOURCE SEPARATION METHODS Yannick DEVILLE Université Paul Sabatier Laboratoire Acoustique, Métrologie, Instrumentation Bât. 3RB2, 8 Route e Narbonne,

More information

Necessary and Sufficient Conditions for Sketched Subspace Clustering

Necessary and Sufficient Conditions for Sketched Subspace Clustering Necessary an Sufficient Conitions for Sketche Subspace Clustering Daniel Pimentel-Alarcón, Laura Balzano 2, Robert Nowak University of Wisconsin-Maison, 2 University of Michigan-Ann Arbor Abstract This

More information

Predictive Control of a Laboratory Time Delay Process Experiment

Predictive Control of a Laboratory Time Delay Process Experiment Print ISSN:3 6; Online ISSN: 367-5357 DOI:0478/itc-03-0005 Preictive Control of a aboratory ime Delay Process Experiment S Enev Key Wors: Moel preictive control; time elay process; experimental results

More information

inflow outflow Part I. Regular tasks for MAE598/494 Task 1

inflow outflow Part I. Regular tasks for MAE598/494 Task 1 MAE 494/598, Fall 2016 Project #1 (Regular tasks = 20 points) Har copy of report is ue at the start of class on the ue ate. The rules on collaboration will be release separately. Please always follow the

More information

WUCHEN LI AND STANLEY OSHER

WUCHEN LI AND STANLEY OSHER CONSTRAINED DYNAMICAL OPTIMAL TRANSPORT AND ITS LAGRANGIAN FORMULATION WUCHEN LI AND STANLEY OSHER Abstract. We propose ynamical optimal transport (OT) problems constraine in a parameterize probability

More information

Multi-View Clustering via Canonical Correlation Analysis

Multi-View Clustering via Canonical Correlation Analysis Kamalika Chauhuri ITA, UC San Diego, 9500 Gilman Drive, La Jolla, CA Sham M. Kakae Karen Livescu Karthik Sriharan Toyota Technological Institute at Chicago, 6045 S. Kenwoo Ave., Chicago, IL kamalika@soe.ucs.eu

More information

Maximal Causes for Non-linear Component Extraction

Maximal Causes for Non-linear Component Extraction Journal of Machine Learning Research 9 (2008) 1227-1267 Submitte 5/07; Revise 11/07; Publishe 6/08 Maximal Causes for Non-linear Component Extraction Jörg Lücke Maneesh Sahani Gatsby Computational Neuroscience

More information

Fast Resampling Weighted v-statistics

Fast Resampling Weighted v-statistics Fast Resampling Weighte v-statistics Chunxiao Zhou Mar O. Hatfiel Clinical Research Center National Institutes of Health Bethesa, MD 20892 chunxiao.zhou@nih.gov Jiseong Par Dept of Math George Mason Univ

More information

A variance decomposition and a Central Limit Theorem for empirical losses associated with resampling designs

A variance decomposition and a Central Limit Theorem for empirical losses associated with resampling designs Mathias Fuchs, Norbert Krautenbacher A variance ecomposition an a Central Limit Theorem for empirical losses associate with resampling esigns Technical Report Number 173, 2014 Department of Statistics

More information

Entanglement is not very useful for estimating multiple phases

Entanglement is not very useful for estimating multiple phases PHYSICAL REVIEW A 70, 032310 (2004) Entanglement is not very useful for estimating multiple phases Manuel A. Ballester* Department of Mathematics, University of Utrecht, Box 80010, 3508 TA Utrecht, The

More information

Relative Entropy and Score Function: New Information Estimation Relationships through Arbitrary Additive Perturbation

Relative Entropy and Score Function: New Information Estimation Relationships through Arbitrary Additive Perturbation Relative Entropy an Score Function: New Information Estimation Relationships through Arbitrary Aitive Perturbation Dongning Guo Department of Electrical Engineering & Computer Science Northwestern University

More information

IN the evolution of the Internet, there have been

IN the evolution of the Internet, there have been 1 Tag-Weighte Topic Moel For Large-scale Semi-Structure Documents Shuangyin Li, Jiefei Li, Guan Huang, Ruiyang Tan, an Rong Pan arxiv:1507.08396v1 [cs.cl] 30 Jul 2015 Abstract To ate, there have been massive

More information

APPROXIMATE SOLUTION FOR TRANSIENT HEAT TRANSFER IN STATIC TURBULENT HE II. B. Baudouy. CEA/Saclay, DSM/DAPNIA/STCM Gif-sur-Yvette Cedex, France

APPROXIMATE SOLUTION FOR TRANSIENT HEAT TRANSFER IN STATIC TURBULENT HE II. B. Baudouy. CEA/Saclay, DSM/DAPNIA/STCM Gif-sur-Yvette Cedex, France APPROXIMAE SOLUION FOR RANSIEN HEA RANSFER IN SAIC URBULEN HE II B. Bauouy CEA/Saclay, DSM/DAPNIA/SCM 91191 Gif-sur-Yvette Ceex, France ABSRAC Analytical solution in one imension of the heat iffusion equation

More information

A simple tranformation of copulas

A simple tranformation of copulas A simple tranformation of copulas V. Durrleman, A. Nikeghbali & T. Roncalli Groupe e Recherche Opérationnelle Créit Lyonnais France July 31, 2000 Abstract We stuy how copulas properties are moifie after

More information

Gaussian processes with monotonicity information

Gaussian processes with monotonicity information Gaussian processes with monotonicity information Anonymous Author Anonymous Author Unknown Institution Unknown Institution Abstract A metho for using monotonicity information in multivariate Gaussian process

More information

A New Discrete Family of Reduced Modified Weibull Distribution

A New Discrete Family of Reduced Modified Weibull Distribution International Journal of Statistical Distributions an Applications 2017; 3(3): 25-31 http://www.sciencepublishinggroup.com/j/ijsa oi: 10.11648/j.ijs.20170303.11 ISSN: 2472-3487 (Print); ISSN: 2472-3509

More information