Prior-Based Dual Additive Latent Dirichlet Allocation for User-Item Connected Documents

Size: px
Start display at page:

Download "Prior-Based Dual Additive Latent Dirichlet Allocation for User-Item Connected Documents"

Transcription

1 Proceeings of the Twenty-Fourth International Joint Conference on Artificial Intelligence (IJCAI 2015) Prior-Base Dual Aitive Latent Dirichlet Allocation for User-Item Connecte Documents Wei Zhang Jianyong Wang Tsinghua National Laboratory for Information Science an Technology (TNList) Department of Computer Science an Technology, Tsinghua University, Beijing, China Abstract User-item connecte ocuments, such as customer reviews for specific items in online shopping website an user tips in location-base social networks, have become more an more prevalent recently. Inferring the topic istributions of user-item connecte ocuments is beneficial for many applications, incluing ocument classification an summarization of users an items. While many ifferent topic moels have been propose for moeling multiple text, most of them cannot account for the ual role of user-item connecte ocuments (each ocument is relate to one user an one item simultaneously) in topic istribution generation process. In this paper, we propose a novel probabilistic topic moel calle Prior-base Dual Aitive Latent Dirichlet Allocation (PDA-). It aresses the ual role of each ocument by associating its Dirichlet prior for topic istribution with user an item topic factors, which leas to a ocument-level asymmetric Dirichlet prior. In the experiments, we evaluate PDA- on several real atasets an the results emonstrate that our moel is effective in comparison to several other moels, incluing hel-out perplexity on moeling text an ocument classification application. 1 Introuction User-item connecte ocuments, which are written by users for specific items, occur in many scenarios. For example, in online shopping websites (e.g., Amazon 1 ), users prefer to write reviews to express their attitues towars some proucts that they have bought. These reviews have become a major reference for caniate buyers to make ecisions. Another example is that in location-base social networks (e.g., Foursquare 2 ), users always propose tips for some point-of-interests (POIs), which are beneficial for other users to gain knowlege about the POIs they have not visite before. In summary, user-item connecte ocuments own ual roles, with each ocument associate with one user an one item simultaneously. They not only reflect the core concerns of users, but also inicate the characteristics of items. Without specific explanation, we also use ocuments to enote user-item connecte ocuments for simplicity in the rest of this paper. Characterizing content of ocuments is an important problem an has many applications in ifferent fiels, e.g., natural language processing an information retrieval. Topic moels [Hofmann, 1999; Blei et al., 2003b] are elegant ways to solve this problem by assuming each ocument has a topic istribution an each topic is represente as a istribution over wors. By this way, ocuments are summarize by their associate topic istributions in a high level. For user-item connecte ocuments, not only the learne topic istributions are useful as usual, but if the topic factors of users an items can be ifferentiate, then they can be use for creating their self-introuction profiles. These profiles are beneficial for other users to gain knowlege about items an retailers to unerstan what their consumers care about. Further, current popular recommener systems can utilize them to make explainable recommenations through content matching. While many topic moels have been propose in the last ecae, almost all of them are esigne for some specific tasks an o not emphasize the ual role phenomenon of useritem connecte ocuments, let alone automatically infer topic factors of users an items simultaneously. For one user-item connecte ocument, its associate topic istribution shoul be influence by its corresponing user an item together, an for a user or an item, its topics consist in multiple ocuments relate to it. Therefore, the main challenge is to connect user, item, an reviews in a unifie topic moel while ensure ifferent combinations of user an item pair ten to generate istinct topic istributions. Besies, it is better for the topic moel to simultaneously infer user an item topic factors. Existing approaches such as author-topic moel [Rosen-Zvi et al., 2004] can only capture user s topic istribution but ignore item s topics. To aress the above issues, we propose a novel generative probabilistic topic moel calle Prior-base Dual Aitive Latent Dirichlet Allocation (PDA-) to moel the useritem connecte ocuments. This moel is somewhat inspire by [Wallach et al., 2009] which points out that asymmetric Dirichlet priors over topic istributions can lea to aitional benefit for topic moels than symmetric priors. In PDA-, we account for the ual role phenomenon by associating Dirichlet prior of each ocument with their corre- 1405

2 sponing user an item. More specifically, PDA- first assumes each user or item is represente with a topic factor. When a component of a topic factor takes a larger value, it means the corresponing user or item concentrates more on that topic. Then an exponential function is utilize to combine a pair of user an topic factor for each ocument through its aitive property of parameters to form a new Dirichlet prior an thus every component of the prior takes positive value. As a result, each ocument has a ifferent Dirichlet prior. The comparison between priors of Latent Dirichlet Allocation () [Blei et al., 2003b], asymmetric Latent Dirichlet Allocation () [Wallach et al., 2009], an PDA- is shown in Table 1. Table 1: Difference of Dirichlet hyper-priors (α, α R K ). PDA- Dir(α), α i = α j (i, j {1,..., K}) Dir(α), α i α j (i, j {1,..., K}) Dir(α )( {1,..., D }) In short, is assigne a corpus-level symmetric prior, AS- has a corpus-level asymmetric prior, while PDA- owns ocument-level asymmetric priors. The main avantages of PDA- lie in three aspects. First of all, it moels text better by accounting for the ual role in topic generations. Secon, it can automatically summarize users an items by the learne topic factors. An last, it constructs ocumentlevel priors with pairs of user an item topic factors. This enables preiction on test ocuments since after moel learning proceure, user an item topic factors are known for these ocuments. Contributions. In all, the main contributions of this paper are summarize as follows: We are the first to aress the ual role phenomenon for topic generations of topic moels in user-item connecte ocuments to the best of our knowlege. We propose a new generative moel calle PDA- which accounts for the ual role phenomenon by connecting user an item topic factors to Dirichlet priors through exponential aitive transformation. We evaluate the propose moel on several real ata sets to verify the effectiveness of PDA-. The experimental results emonstrate that our moel not only achieves superior hel-out perplexity on test ata, but generates better topics for its goo performance in ocument classification application as well. In what follows, etaile moel specification is introuce in Section 2. Then we provie experimental results an some analysis in Section 3. Relate works are briefly iscusse in Section 4. In the last, we raw a conclusion for this work.. 2 Approach This section will be ivie into three parts. The first part is mainly about the escription of PDA-, incluing the generative process of the moel. Then we erive the Gibbs EM base learning algorithm for the moel. Finally, a concise introuction to preiction on test ata is given. Before we Figure 1: Graphical moel of PDA-. procee, the main mathematical symbols we will use later are shown for reference in Table Moel Description Stanar topic moels such as always assume each ocument relates to a istribution over K latent components calle topics an each topic can be explaine by a istribution over all wors in vocabulary W. Both the topic an wor istributions are sample from a Dirichlet istribution Dir(α) known as prior in Bayesian learning. Our moel follows the basic topic generation process of. The key ifference lies in how we set the parameter α of Dirichlet prior for each useritem connecte ocument. The hyper-parameter α has been emonstrate to influence the result of text moeling [Wallach et al., 2009]. As we iscusse, each user or item has its own characteristics. We assume each user is associate with a topic factor φ u R K an the same for each item, i.e., φ i R K. In topic factors, components with larger value enote users an items concentrate more on those corresponing topics. Topic factors shoul influence generation of topic istributions for ocuments. To brige the gap between them, one natural iea is to associate topic factors with Dirichlet hyper-parameter α as in Bayesian learning framework, prior istribution reveals prior knowlege about how ata will be generate before it really comes into action. This is consistent with the intrinsics of generation of user-item connecte ocuments. Imagine that when a user plans to write something about an item, the first thing he will o is to choose some themes from what he often concentrates on an also conforme to the context of the item. Obviously, the themes are both relevant to the user an the item. These themes can be regare as latent topics stuie in topic moels. For a ocument u,v, we associate user an item topic factors with Dirichlet prior parameter of its topic istribution as follows, = exp(φ u + φ v ) (1) where exponential transformation ensures each component of to be positive, which satisfies the requirements of Dirichlet parameters. So far, topic factors of user an item can 1406

3 Table 2: Notations of main aopte symbols. D W u U v V K λ U λ V λ B φ u φ v φ b α θ η β k z j w j u,v N u,v N u,v,k Mw D Mw,k D User-item connecte ocument set Vocabulary set An user in the user set U An item in the item set V Number of topics Precision of Gaussian prior for user factor φ u Precision of Gaussian prior for item factor φ v Precision of Gaussian prior for backgroun factor φ b Topic factor of user u Topic factor of item v Topic factor of backgroun Dirichlet prior parameter for topic istribution θ Topic istribution of ocument Dirichlet prior parameter for wor istribution β Multinomial istribution over wors of topic k Latent topic assignment for wor w j The j-th wor in ocument The ocument written by user u for item v Number of wors in ocument u,v Number of wors with topic k in ocument Occurrence of wor w in corpus Occurrence of wor w with topic k in corpus influence the generation of topic istribution of ocument u,v through. One natural extension is to consier backgroun topic factor φ b aitionally. It is necessary since some wors such as conjunction wors frequently occur in ocuments. The topic relates to these wors are not specially owne by users or items. Therefore, the complete transformation between topic factors an Dirichlet parameter is efine to be = exp(φ u + φ v + φ b ) (2) Base on the above analysis, we can summarize the generative story of PDA- for user-item connecte ocuments as below, 1. Draw backgroun topic factor φ b N (λ B ). 2. For each user u, raw user topic factor φ u N (λ U ). 3. For each item v, raw item topic factor φ v N (λ V ). 4. For each topic k, raw wor istribution β k Dirichlet(η). 5. For each user-item connecte ocument u,v in ocument collection D, (a) Draw topic istribution θ Dirichlet( ) where is calculate through Equation(2). (b) For each position j in the ocument u,v : (b.1) Sample topic z j Mult(θ u,v). (b.2) Draw wor w j Mult(β zj ). The complete graphical moel representation of our propose moel is provie in Figure 1. The joint probability of a whole corpus D containing user-item connecte ocuments an topic factors Θ t = {φ U, φ V, φ b } is efine as follows, P (D, Θ t) = N (φ b ; 0, λ B) u K Dir(β k ; η) k =1 j N u,v z j N (φ u; 0, λ U ) v D u,v =1 N (φ v; 0, λ V ) Dir(θ u,v ; αu,v ) P (z j θ u,v )P (wj βz j )θu,v β k (3) 2.2 Learning In this work, our goal is to learn optimal topic factors Θ t an key latent variables θ an β by maximizing the log-likelihoo of joint probability shown in Equation (3), L = log ( P (D, Θ t ) ) (4) However, it is intractable to irectly optimize the above function an compute the posterior istribution of θ an β ue to the summarization of latent topics z i in iscrete space an coupling of θ an β. Luckily, if the assignments of latent topics can be inferre for all wors in reviews, then not only θ an β can be calculate, but also Θ t are tractable to be optimize. This intuition leas to the iea of Gibb EM algorithm. Gibbs EM learning algorithm is wiely aopte in (partially) Bayesian latent factor moels [Wallach, 2006; Liu et al., 2013] which alternates between sampling the value of latent variables (Expectation Step) an optimizing some moel parameters (Maximization Step). More specifically, we aopt collapse Gibbs Sampling to etermine all the latent topics in E-step an optimize Θ t through graient ascent in M-step. E-step Given a wor position j in ocument u,v, the key inferential problem in collapse Gibbs sampling is to erive the posterior istribution of its associate topic, i.e., P (z u,v k z u,v, j, wu,v,j = ) where zu,v, j enotes all topic assignments of wors in ocument are known except the wor in position j. As Θ t are fixe in E-step, can be calculate through Equation (2). By applying Bayesian formula an conitional epenence property, it is easy to erive the following formal efinition of the posterior istribution accoring to the results of [Griffiths an Steyvers, 2004], P (z u,v,j = k zu,v, j, wu,v ) = N u,v,k + αu,v,k N + K,k M D w j,k + η Mw D + W η (5) where all the counts o not inclue the current wor. The above equation is very similar to that in stanar topic moel except the Dirichlet prior,k. M-step After getting the latent topic assignments for all wors in reviews, the objective log-likelihoo of joint probability becomes P (D, Z, Θ t ) an thus the summarization term in Equation (3) vanishes. Due to the conjugate prior property of Dirichlet multinomial istribution [Heinrich, 2004; Griffiths an Steyvers, 2004], θ an β can be integrate out an P (D, Z, Θ t ) is now reformulate as below, P (D, Z, Θ t ) = N (φ b ; 0, λ B ) u K u,v k=1 Γ(N,k + αu,v,k ) Γ(N u,v u,v K k=1 u,v w Γ(N N (φ u ; 0, λ U ) N (φ v ; 0, λ V ) v + Γ( K k=1 αu,v,k ) K k=1 αu,v,k ) K k=1 Γ(αu,v,k ),k,w + η) Γ(N u,v Γ( W η),k + W η) w Γ(η) (6) 1407

4 where Γ is Gamma function with form Γ(x) = (x 1)! when x is an integer. By maximizing the log-likelihoo of the above joint probability with graient ascent algorithm, we can get optimal Θ t. Graient ascent consists of two steps. The first step is to euce the graients of the parameters. For example, the graient of φ u,k is L φ u,k = u u,v,k + Ψ(N u,v λ U φ u,k,k + αu,v,k ( Ψ( k,k ) Ψ(,k ) u,v ) Ψ(N + ),k ) k where Ψ is Digamma function which is equal to logarithmic erivative of the Gamma function. The graients of φ v,k an φ b,k are analogous to φ u,k except the summarization conition an regularization term. We shoul emphasize although for one review, φ u,k an φ v,k are similar, ifferent reviews connect to ifferent users an items an thus the upates for the users an items are istinct from a whole perspective. Base on the obtaine graients, the secon step of graient ascent is to upate moel parameters through (7) Θ t+1 = Θ t + ω L Θ t (8) where ω is the learning rate of the algorithm. Superscript t enotes the finishe number of iterations. Besies, after the iterative learning process converges, we can calculate β an θ by collecting enough subsequent samples though a burn-in process as [Blei et al., 2003a], t β k,w = M t,w,k D + η t M w D (9) + W η θ,k = t N u,v t,,k + αu,v,k t N + k αu,v,k (10) where N u,v t,,k enotes the samples collecte in the t iteration. In summary, the whole learning algorithm for PDA- is conclue in Algorithm Preiction When the PDA- moel is utilize for ocument moeling on test ataset, Θ t an β shoul be fixe to be the optimal values learne from training set. The only important step is to infer latent topic assignments over test ocuments by sampling from a resembling formula as Equation (5), which aitionally incorporates relevant count variables of test ocuments. 3 Experiments In this section, we first introuce the atasets we use an the preprocessing steps for them. Then we iscuss the aopte comparison moels an the hyper-parameter setting of PDA-. Finally, we analyze the experimental results to emonstrate the effectiveness of our metho. Algorithm 1: The Gibbs EM Algorithm for PDA- Input: User-item connect ocument collection D. Output: Optimal user, item, an backgroun topic factors Θ t, ocument-topic istribution θ, topic-wor istribution β 1 Initialize hyper-parameters an Θ t 2 Ranomly raw latent topic assignments of wors in ocuments 3 tt = 0 4 while Not converge an tt Iter max o 5 E-step: 6 (a) Sampling all wors latent topic assignments 7 through Equation (5) 8 M-step: 9 (a) Calculate the graients through Equation (7) 10 (b) Upate Θ t through Equation (8) 11 tt tt Calculate θ an β through Equation (9) an (10) uner a burn-in process 3.1 Dataset We aopt three real ata collections from Yelp 3 an [McAuley an Leskovec, 2013]. Base on their origins, we enominate the three ata sets as Yelp, AmazonFoo an AmazonSport, respectively. To clean text ata, we aopt the following four processing steps: (1) converting all letters into lowercase, (2) filtering stop wors 4 an punctuations, (3) removing reviews which are too short, an (4) saving frequent wors to form a vocabulary. After cleaning text ata, we further remove users an items with less than 5 reviews to ensure that users an items have enough relate ocuments. Finally, we obtain the experimental ata sets whose basic statistics are shown in Table 3. We ranomly ivie the two collections into train, valiation, an test set with the ratio 7 : 1 : 2 for testing hel-out perplexity an further binary ocument classification task. Table 3: Introuction of Experimental Datasets. Data User Item Reviews Length Yelp (wors) AmazonFoo (wors) AmazonSport (wors) 3.2 Comparison Moels To verify the effectiveness of our moel, we first introuce the comparison moels aopte in the experiments. Latent Dirichlet Allocation (): Although [Blei et al., 2003b] can neither capture author topics, nor obtain item topics, it is still necessary to analyze its result to verify the effectiveness of an PDA- for exploring asymmetric Dirichlet prior as we mentione in Section challenge

5 Perplexity Dimension PDA- (a) Change Perplexity on Yelp. Perplexity Dimension PDA- (b) Change Perplexity on AmazonFoo. Perplexity Dimension PDA- (c) Change Perplexity on AmazonSport. Figure 2: Result in terms of Perplexity with change of topic components. Author-topic Moel (ATM): We aapt ATM [Rosen-Zvi et al., 2004] to our problem by regaring users as authors in the moel. ATM somewhat resembles PDA- as both of them can obtain users topics. However, ATM cannot hanle the ouble role phenomenon in user-item connecte ocuments. Intuitively, it will perform worse than the new moel. Gibbs Sampler for Dirichlet Multinomial (GSDMM): GSDMM [Yin an Wang, 2014] is a stanar topic moel for hanling short text such as tweets in social meia by assuming all wors in a ocument have the same topic an use as a comparison here ue to the short length of the ocuments shown in Table 3, Asymmetric Latent Dirichlet Allocation (): AS- [Wallach et al., 2009] is a strong comparison as it aresses popularity bias for each topic by incorporating asymmetric Dirichlet prior. It can be regare as the most similar metho to PDA- which only consiers the backgroun topic factor for topic generation process. 3.3 Parameter Setting All the hyper-parameters are etermine base on their performances on valiation atasets. For all the comparison methos an PDA-, we assign 0.1 to η. For comparisons except, we choose α to be 0.1 as well for its goo performance. The concentration parameter α in is tunne to be 0.1. Apart from η, λ U, λ V, an λ b are set to be 1 for PDA- uniformly. 3.4 Results Results on Text Moeling We first compare all the aopte moels in terms of perplexity on test atasets. Perplexity is wiely use in probabilistic moel for checking their quality. It is irectly relevant to log-likelihoo of probability an normally efine as below, Perplexity(D test ) = exp ( D test log P () D test N ) (11) The etaile results are shown in Table 4. We notice that GSDMM performs not very well in user-item connecte ocument moeling although the average length of the ocuments is short. One reason is that user-item connecte ocuments are still longer than tweets in social meia an very short text is more likely to contain only one topic. Table 4: Comparisons of ifferent moels in terms of Perplexity on test ata with K = 40. Metho Yelp AmazonFoo AmazonSport ATM GSDMM PDA ATM ose not show promising results although it moels the role of users in topic generation. This may be explaine by the fact that ATM is originally esigne for the scenario where each ocument has multiple authors (e.g., research papers). As each user-item connecte ocument is uniquely owne by one user, the ATM may not hanle it very well. outperforms consistently in three atasets. This makes sense that asymmetric Dirichlet prior captures popularity bias of topics in generation process from a corpus level while ignores the bias. It emonstrates that the iea of moifying hyper-parameters of Dirichlet prior for topic istribution is promising an also provies evience for PDA- to further ifferentiate prior parameters. It is clear that PDA- achieves superior performances among all the aopte moels. Especially, PDA- behaves better than the strong comparison significantly. This inicates that asymmetric Dirichlet prior which capture topic concentration bias from more graine ocument level is better than from corpus level. Moreover, we iscover gains a minor improvement over in the last ataset, while the ecrease of perplexity of PDA- is prominent, which inicates our moel has wier applicability. Dimension Influence on Text Moeling We analyze the performance evolution with change of number of latent components from 10 to 100 with a step size of 10. We only compare,, an PDA- as their results are closer than the other two moels. As Figure 2 shown, the perplexity of all the three moels ecrease steaily with increase of topic components. We choose K = 40 for other experiments in this work by consiering a trae-off between efficiency an effectiveness as larger number of components costs more time to learn moels. Besies, the perplexity if- 1409

6 Table 5: Comparisons of ifferent methos in terms of Accuracy, Precision, an F1 Metric on classification task with K=40. Metho Yelp AmazonFoo AmazonSport Accuracy Precision F1 Accuracy Precision F1 Accuracy Precison F PDA ferences between these moels are relatively evient when K 40 an thus the comparisons in experiments will not be influence. In general, PDA- outperforms an regarless of how the number of topic components varies. Application to Document Classification As the perplexity oes not correlate with human jugments, it may be better to test topic moels when using its generate topics for other tasks [Chang et al., 2009]. In this work, we apply,, an PDA- to binary ocument classification task by utilizing topic istributions learne from them as features for supervise classification moels. We shoul emphasize our goal here is to test the quality of learne topics, but not to compare with state-of-the-art ocument classification methos. On the other han, features of topic moels can be regare as a complementary for other supervise methos. Base on the rating scores associate with the three atasets, we ivie all the ocuments of training, valiation, an test sets into two classes. The critical value between for all atasets is specifie to be 3.5 ([1-5]). The number of ocuments in each class are shown in Table 6. Table 6: Statistics of classification atasets. Data 1st-class 2n-class Yelp AmazonFoo AmazonSport We compare,, an PDA- in terms of average accuracy, precision, an F1 metrics. Among them, precision an F1 are first compute in each class an then take weighte average over two classes. We aopt stanar supervise classifier, i.e., Support Vector Machine (SVM), in the experiments. As the results shown in Table 5, PDA- achieves best performance across all the atasets. Hence the quality of topics generate by PDA- turns out to be better than those of the other moels. Besies, although gains lower perplexity than in Yelp ataset, it oes not perform better in terms of classification metrics. This reveals obtaining lower perplexity oes not ensure better iscriminative power of features for classification performance efinitely, which also reveals the robustness of PDA-. Case Stuy To give an intuitive interpretation of PDA-, we ranomly sample two users an two items to reveal their topic factors in Figure 3, an show their representative topics (topic inex correspons to the largest value for each user an item topic factor) learne from AmazonFoo ata in Table 7. We fin that the sample users an items only focus on several topics, for many values less than 0. Thus, they can be characterize by a few critical topics which have clear inications. For example, user-2 likes chips very much. Weight Topic Inex USER-1 USER-2 ITEM-1 ITEM-2 Figure 3: Sample user an item topic factors. Table 7: Selecte topics from AmazonFoo ata set. T3: amazon, price, store, buy, local, orer, prouct, fin, bought, shipping T9: foo, cat, cats, chicken, eat, tuna, canne, fish, cans, flavors T25: chips, flavor, salt, potato, sweet, bag, taste, chip, popchips, goo T27: bars, bar, fat, calories, snack, fiber, grams, protein, healthy, sugar 4 Relate Work Topic moels are popularly utilize for moeling text. Different topic moels concentrate on ifferent aspects of text. For example, [Blei an McAuliffe, 2007] exploite the labels corresponing to ocuments to construct their supervise topic moel. The authors in [Mei et al., 2008] consiere the link relations between ocuments an assume linke ocuments have similar topic istributions. Recently, many topic moels have been propose for moeling review text [Titov an McDonal, 2008; Zhao et al., 2010; Sauper et al., 2011; Moghaam an Ester, 2011; Mukherjee et al., 2014]. Many of them are esigne for sentiment analysis, incluing aspects mining an rating preiction. While our moel is not esigne specially for review text, but a wier range of text, i.e., user-item connecte ocuments an we concentrate on accounting for the ual role in the topic generation process of the text which are ignore in these moels. Collaborative topic moel (CTM) [Wang an Blei, 2011] consiere the user an item information simultaneously for preicting ratings of users for items. Nevertheless, its topic generation process is the similar to, which oes not take the ual role into consieration. Author-topic moel [Rosen-Zvi et al., 2004] is somewhat similar with our moel as it can obtain the topic 1410

7 istribution of users. However, the topic generation is only influence by users an it still cannot moel the ual role phenomenon. Many papers eal with ual roles in ifferent research fiels, such as the ual role of hashtag in social networks [Yang et al., 2012] an users in question answering [Xu et al., 2012]. However, few previous works emphasize the ual role phenomenon in topic generation process of topic moels. 5 Conclusion In this paper, we propose a new probabilistic topic moel calle PDA- to account for the ual role phenomenon of user-item connecte ocuments. PDA- moels the topic generation process through the joint effect of user, item, an backgroun topic factors on the Dirichlet prior of topic istribution. A Gibbs EM base learning algorithm is erive for the new moel to learn optimal topic factors. The experimental results on real ata collections have shown that PDA- achieves better hel-out perplexity an binary classification accuracy than several other moels. Acknowlegments This work was supporte in part by National Basic Research Program of China (973 Program) uner Grant No. 2011CB302206, National Natural Science Founation of China uner Grant No , an Tsinghua University Initiative Scientific Research Program. References [Blei an McAuliffe, 2007] Davi M. Blei an Jon D. McAuliffe. Supervise topic moels. In in Proc. of NIPS 07, Vancouver, British Columbia, Canaa, December 3-6, 2007, pages , [Blei et al., 2003a] Davi M. Blei, Thomas L. Griffiths, Michael I. Joran, an Joshua B. Tenenbaum. Hierarchical topic moels an the neste chinese restaurant process. In in Proc. of NIPS 03 December 8-13, 2003, Vancouver an Whistler, British Columbia, Canaa], pages 17 24, [Blei et al., 2003b] Davi M. Blei, Anrew Y. Ng, an Michael I. Joran. Latent irichlet allocation. Journal of Machine Learning Research, 3: , [Chang et al., 2009] Jonathan Chang, Joran L. Boy-Graber, Sean Gerrish, Chong Wang, an Davi M. Blei. Reaing tea leaves: How humans interpret topic moels. In in Proc. of NIPS 09. Proceeings of a meeting hel 7-10 December 2009, Vancouver, British Columbia, Canaa., pages , [Griffiths an Steyvers, 2004] Thomas L. Griffiths an Mark Steyvers. Fining scientific topics. PNAS, 101(suppl. 1): , [Heinrich, 2004] Gregor Heinrich. Parameter estimation for text analysis. Technical report, [Hofmann, 1999] Thomas Hofmann. Probabilistic latent semantic inexing. In in Proc. of SIGIR 99, August 15-19, 1999, Berkeley, CA, USA, pages 50 57, [Liu et al., 2013] Bin Liu, Yanjie Fu, Zijun Yao, an Hui Xiong. Learning geographical preferences for point-of-interest recommenation. in Proc. of KDD 13, pages , New York, NY, USA, ACM. [McAuley an Leskovec, 2013] Julian J. McAuley an Jure Leskovec. Hien factors an hien topics: unerstaning rating imensions with review text. In in Proc. of RecSys 13, Hong Kong, China, October 12-16, 2013, pages , [Mei et al., 2008] Qiaozhu Mei, Deng Cai, Duo Zhang, an ChengXiang Zhai. Topic moeling with network regularization. In in Proc. of WWW 08, Beijing, China, April 21-25, 2008, pages , [Moghaam an Ester, 2011] Samaneh Moghaam an Martin Ester. I: interepenent moel for learning latent aspects an their ratings from online prouct reviews. In in Proc. of SIGIR 11, Beijing, China, July 25-29, 2011, pages , [Mukherjee et al., 2014] Subhabrata Mukherjee, Gaurab Basu, an Sachinra Joshi. Joint author sentiment topic moel. In in Proc. of SDM 14, Philaelphia, Pennsylvania, USA, April 24-26, 2014, pages , [Rosen-Zvi et al., 2004] Michal Rosen-Zvi, Thomas L. Griffiths, Mark Steyvers, an Pahraic Smyth. The author-topic moel for authors an ocuments. In in Proc. of UAI 04, Banff, Canaa, July 7-11, 2004, pages , [Sauper et al., 2011] Christina Sauper, Aria Haghighi, an Regina Barzilay. Content moels with attitue. In in Proc. of ACL 11, June, 2011, Portlan, Oregon, USA, pages , [Titov an McDonal, 2008] Ivan Titov an Ryan T. McDonal. Moeling online reviews with multi-grain topic moels. In in Proc. of WWW 08, Beijing, China, April 21-25, 2008, pages , [Wallach et al., 2009] Hanna M. Wallach, Davi M. Mimno, an Anrew McCallum. Rethinking : why priors matter. In in Proc. of NIPS 09. December 7-10, 2009, Vancouver, British Columbia, Canaa., pages , [Wallach, 2006] Hanna M. Wallach. Topic moeling: Beyon bagof-wors. ICML 06, pages , New York, NY, USA, ACM. [Wang an Blei, 2011] Chong Wang an Davi M. Blei. Collaborative topic moeling for recommening scientific articles. In Proceeings of the 17th ACM SIGKDD International Conference on Knowlege Discovery an Data Mining, San Diego, CA, USA, August 21-24, 2011, pages , [Xu et al., 2012] Fei Xu, Zongcheng Ji, an Bin Wang. Dual role moel for question recommenation in community question answering. In in Proc. of SIGIR 12, Portlan, OR, USA, August 12-16, 2012, pages , [Yang et al., 2012] Lei Yang, Tao Sun, Ming Zhang, an Qiaozhu Mei. We know #tag: oes the ual role affect hashtag aoption? In in Proc. of WWW 12, Lyon, France, April 16-20, 2012, pages , [Yin an Wang, 2014] Jianhua Yin an Jianyong Wang. A irichlet multinomial mixture moel-base approach for short text clustering. In in Proc. of KDD 14, New York, NY, USA - August 24-27, 2014, pages , [Zhao et al., 2010] Wayne Xin Zhao, Jing Jiang, Hongfei Yan, an Xiaoming Li. Jointly moeling aspects an opinions with a maxent-la hybri. In in Proc. of EMNLP 10, 9-11 October 2010, MIT Stata Center, Massachusetts, USA, A meeting of SIG- DAT, a Special Interest Group of the ACL, pages 56 65,

Collapsed Gibbs and Variational Methods for LDA. Example Collapsed MoG Sampling

Collapsed Gibbs and Variational Methods for LDA. Example Collapsed MoG Sampling Case Stuy : Document Retrieval Collapse Gibbs an Variational Methos for LDA Machine Learning/Statistics for Big Data CSE599C/STAT59, University of Washington Emily Fox 0 Emily Fox February 7 th, 0 Example

More information

Survey Sampling. 1 Design-based Inference. Kosuke Imai Department of Politics, Princeton University. February 19, 2013

Survey Sampling. 1 Design-based Inference. Kosuke Imai Department of Politics, Princeton University. February 19, 2013 Survey Sampling Kosuke Imai Department of Politics, Princeton University February 19, 2013 Survey sampling is one of the most commonly use ata collection methos for social scientists. We begin by escribing

More information

LDA Collapsed Gibbs Sampler, VariaNonal Inference. Task 3: Mixed Membership Models. Case Study 5: Mixed Membership Modeling

LDA Collapsed Gibbs Sampler, VariaNonal Inference. Task 3: Mixed Membership Models. Case Study 5: Mixed Membership Modeling Case Stuy 5: Mixe Membership Moeling LDA Collapse Gibbs Sampler, VariaNonal Inference Machine Learning for Big Data CSE547/STAT548, University of Washington Emily Fox May 8 th, 05 Emily Fox 05 Task : Mixe

More information

Topic Modeling: Beyond Bag-of-Words

Topic Modeling: Beyond Bag-of-Words Hanna M. Wallach Cavenish Laboratory, University of Cambrige, Cambrige CB3 0HE, UK hmw26@cam.ac.u Abstract Some moels of textual corpora employ text generation methos involving n-gram statistics, while

More information

. Using a multinomial model gives us the following equation for P d. , with respect to same length term sequences.

. Using a multinomial model gives us the following equation for P d. , with respect to same length term sequences. S 63 Lecture 8 2/2/26 Lecturer Lillian Lee Scribes Peter Babinski, Davi Lin Basic Language Moeling Approach I. Special ase of LM-base Approach a. Recap of Formulas an Terms b. Fixing θ? c. About that Multinomial

More information

CUSTOMER REVIEW FEATURE EXTRACTION Heng Ren, Jingye Wang, and Tony Wu

CUSTOMER REVIEW FEATURE EXTRACTION Heng Ren, Jingye Wang, and Tony Wu CUSTOMER REVIEW FEATURE EXTRACTION Heng Ren, Jingye Wang, an Tony Wu Abstract Popular proucts often have thousans of reviews that contain far too much information for customers to igest. Our goal for the

More information

Introduction to Machine Learning

Introduction to Machine Learning How o you estimate p(y x)? Outline Contents Introuction to Machine Learning Logistic Regression Varun Chanola April 9, 207 Generative vs. Discriminative Classifiers 2 Logistic Regression 2 3 Logistic Regression

More information

Latent Dirichlet Allocation in Web Spam Filtering

Latent Dirichlet Allocation in Web Spam Filtering Latent Dirichlet Allocation in Web Spam Filtering István Bíró Jácint Szabó Anrás A. Benczúr Data Mining an Web search Research Group, Informatics Laboratory Computer an Automation Research Institute of

More information

The new concepts of measurement error s regularities and effect characteristics

The new concepts of measurement error s regularities and effect characteristics The new concepts of measurement error s regularities an effect characteristics Ye Xiaoming[1,] Liu Haibo [3,,] Ling Mo[3] Xiao Xuebin [5] [1] School of Geoesy an Geomatics, Wuhan University, Wuhan, Hubei,

More information

Collapsed Variational Inference for HDP

Collapsed Variational Inference for HDP Collapse Variational Inference for HDP Yee W. Teh Davi Newman an Max Welling Publishe on NIPS 2007 Discussion le by Iulian Pruteanu Outline Introuction Hierarchical Bayesian moel for LDA Collapse VB inference

More information

Computing Exact Confidence Coefficients of Simultaneous Confidence Intervals for Multinomial Proportions and their Functions

Computing Exact Confidence Coefficients of Simultaneous Confidence Intervals for Multinomial Proportions and their Functions Working Paper 2013:5 Department of Statistics Computing Exact Confience Coefficients of Simultaneous Confience Intervals for Multinomial Proportions an their Functions Shaobo Jin Working Paper 2013:5

More information

Schrödinger s equation.

Schrödinger s equation. Physics 342 Lecture 5 Schröinger s Equation Lecture 5 Physics 342 Quantum Mechanics I Wenesay, February 3r, 2010 Toay we iscuss Schröinger s equation an show that it supports the basic interpretation of

More information

Lecture 2: Correlated Topic Model

Lecture 2: Correlated Topic Model Probabilistic Moels for Unsupervise Learning Spring 203 Lecture 2: Correlate Topic Moel Inference for Correlate Topic Moel Yuan Yuan First of all, let us make some claims about the parameters an variables

More information

Level Construction of Decision Trees in a Partition-based Framework for Classification

Level Construction of Decision Trees in a Partition-based Framework for Classification Level Construction of Decision Trees in a Partition-base Framework for Classification Y.Y. Yao, Y. Zhao an J.T. Yao Department of Computer Science, University of Regina Regina, Saskatchewan, Canaa S4S

More information

JUST THE MATHS UNIT NUMBER DIFFERENTIATION 2 (Rates of change) A.J.Hobson

JUST THE MATHS UNIT NUMBER DIFFERENTIATION 2 (Rates of change) A.J.Hobson JUST THE MATHS UNIT NUMBER 10.2 DIFFERENTIATION 2 (Rates of change) by A.J.Hobson 10.2.1 Introuction 10.2.2 Average rates of change 10.2.3 Instantaneous rates of change 10.2.4 Derivatives 10.2.5 Exercises

More information

Cascaded redundancy reduction

Cascaded redundancy reduction Network: Comput. Neural Syst. 9 (1998) 73 84. Printe in the UK PII: S0954-898X(98)88342-5 Cascae reunancy reuction Virginia R e Sa an Geoffrey E Hinton Department of Computer Science, University of Toronto,

More information

Homework 2 Solutions EM, Mixture Models, PCA, Dualitys

Homework 2 Solutions EM, Mixture Models, PCA, Dualitys Homewor Solutions EM, Mixture Moels, PCA, Dualitys CMU 0-75: Machine Learning Fall 05 http://www.cs.cmu.eu/~bapoczos/classes/ml075_05fall/ OUT: Oct 5, 05 DUE: Oct 9, 05, 0:0 AM An EM algorithm for a Mixture

More information

A Course in Machine Learning

A Course in Machine Learning A Course in Machine Learning Hal Daumé III 12 EFFICIENT LEARNING So far, our focus has been on moels of learning an basic algorithms for those moels. We have not place much emphasis on how to learn quickly.

More information

Information retrieval LSI, plsi and LDA. Jian-Yun Nie

Information retrieval LSI, plsi and LDA. Jian-Yun Nie Information retrieval LSI, plsi and LDA Jian-Yun Nie Basics: Eigenvector, Eigenvalue Ref: http://en.wikipedia.org/wiki/eigenvector For a square matrix A: Ax = λx where x is a vector (eigenvector), and

More information

IN the evolution of the Internet, there have been

IN the evolution of the Internet, there have been 1 Tag-Weighte Topic Moel For Large-scale Semi-Structure Documents Shuangyin Li, Jiefei Li, Guan Huang, Ruiyang Tan, an Rong Pan arxiv:1507.08396v1 [cs.cl] 30 Jul 2015 Abstract To ate, there have been massive

More information

Admin BACKPROPAGATION. Neural network. Neural network 11/3/16. Assignment 7. Assignment 8 Goals today. David Kauchak CS158 Fall 2016

Admin BACKPROPAGATION. Neural network. Neural network 11/3/16. Assignment 7. Assignment 8 Goals today. David Kauchak CS158 Fall 2016 Amin Assignment 7 Assignment 8 Goals toay BACKPROPAGATION Davi Kauchak CS58 Fall 206 Neural network Neural network inputs inputs some inputs are provie/ entere Iniviual perceptrons/ neurons Neural network

More information

Least-Squares Regression on Sparse Spaces

Least-Squares Regression on Sparse Spaces Least-Squares Regression on Sparse Spaces Yuri Grinberg, Mahi Milani Far, Joelle Pineau School of Computer Science McGill University Montreal, Canaa {ygrinb,mmilan1,jpineau}@cs.mcgill.ca 1 Introuction

More information

This module is part of the. Memobust Handbook. on Methodology of Modern Business Statistics

This module is part of the. Memobust Handbook. on Methodology of Modern Business Statistics This moule is part of the Memobust Hanbook on Methoology of Moern Business Statistics 26 March 2014 Metho: Balance Sampling for Multi-Way Stratification Contents General section... 3 1. Summary... 3 2.

More information

A Review of Multiple Try MCMC algorithms for Signal Processing

A Review of Multiple Try MCMC algorithms for Signal Processing A Review of Multiple Try MCMC algorithms for Signal Processing Luca Martino Image Processing Lab., Universitat e València (Spain) Universia Carlos III e Mari, Leganes (Spain) Abstract Many applications

More information

7.1 Support Vector Machine

7.1 Support Vector Machine 67577 Intro. to Machine Learning Fall semester, 006/7 Lecture 7: Support Vector Machines an Kernel Functions II Lecturer: Amnon Shashua Scribe: Amnon Shashua 7. Support Vector Machine We return now to

More information

05 The Continuum Limit and the Wave Equation

05 The Continuum Limit and the Wave Equation Utah State University DigitalCommons@USU Founations of Wave Phenomena Physics, Department of 1-1-2004 05 The Continuum Limit an the Wave Equation Charles G. Torre Department of Physics, Utah State University,

More information

Speaker Adaptation Based on Sparse and Low-rank Eigenphone Matrix Estimation

Speaker Adaptation Based on Sparse and Low-rank Eigenphone Matrix Estimation INTERSPEECH 2014 Speaker Aaptation Base on Sparse an Low-rank Eigenphone Matrix Estimation Wen-Lin Zhang 1, Dan Qu 1, Wei-Qiang Zhang 2, Bi-Cheng Li 1 1 Zhengzhou Information Science an Technology Institute,

More information

Content-based Recommendation

Content-based Recommendation Content-based Recommendation Suthee Chaidaroon June 13, 2016 Contents 1 Introduction 1 1.1 Matrix Factorization......................... 2 2 slda 2 2.1 Model................................. 3 3 flda 3

More information

'HVLJQ &RQVLGHUDWLRQ LQ 0DWHULDO 6HOHFWLRQ 'HVLJQ 6HQVLWLYLW\,1752'8&7,21

'HVLJQ &RQVLGHUDWLRQ LQ 0DWHULDO 6HOHFWLRQ 'HVLJQ 6HQVLWLYLW\,1752'8&7,21 Large amping in a structural material may be either esirable or unesirable, epening on the engineering application at han. For example, amping is a esirable property to the esigner concerne with limiting

More information

The Press-Schechter mass function

The Press-Schechter mass function The Press-Schechter mass function To state the obvious: It is important to relate our theories to what we can observe. We have looke at linear perturbation theory, an we have consiere a simple moel for

More information

Optimization of Geometries by Energy Minimization

Optimization of Geometries by Energy Minimization Optimization of Geometries by Energy Minimization by Tracy P. Hamilton Department of Chemistry University of Alabama at Birmingham Birmingham, AL 3594-140 hamilton@uab.eu Copyright Tracy P. Hamilton, 1997.

More information

Influence of weight initialization on multilayer perceptron performance

Influence of weight initialization on multilayer perceptron performance Influence of weight initialization on multilayer perceptron performance M. Karouia (1,2) T. Denœux (1) R. Lengellé (1) (1) Université e Compiègne U.R.A. CNRS 817 Heuiasyc BP 649 - F-66 Compiègne ceex -

More information

Vectors in two dimensions

Vectors in two dimensions Vectors in two imensions Until now, we have been working in one imension only The main reason for this is to become familiar with the main physical ieas like Newton s secon law, without the aitional complication

More information

Pachinko Allocation: DAG-Structured Mixture Models of Topic Correlations

Pachinko Allocation: DAG-Structured Mixture Models of Topic Correlations : DAG-Structured Mixture Models of Topic Correlations Wei Li and Andrew McCallum University of Massachusetts, Dept. of Computer Science {weili,mccallum}@cs.umass.edu Abstract Latent Dirichlet allocation

More information

18 EVEN MORE CALCULUS

18 EVEN MORE CALCULUS 8 EVEN MORE CALCULUS Chapter 8 Even More Calculus Objectives After stuing this chapter you shoul be able to ifferentiate an integrate basic trigonometric functions; unerstan how to calculate rates of change;

More information

Part I: Web Structure Mining Chapter 1: Information Retrieval and Web Search

Part I: Web Structure Mining Chapter 1: Information Retrieval and Web Search Part I: Web Structure Mining Chapter : Information Retrieval an Web Search The Web Challenges Crawling the Web Inexing an Keywor Search Evaluating Search Quality Similarity Search The Web Challenges Tim

More information

Estimation of the Maximum Domination Value in Multi-Dimensional Data Sets

Estimation of the Maximum Domination Value in Multi-Dimensional Data Sets Proceeings of the 4th East-European Conference on Avances in Databases an Information Systems ADBIS) 200 Estimation of the Maximum Domination Value in Multi-Dimensional Data Sets Eleftherios Tiakas, Apostolos.

More information

Homework 2 EM, Mixture Models, PCA, Dualitys

Homework 2 EM, Mixture Models, PCA, Dualitys Homework 2 EM, Mixture Moels, PCA, Dualitys CMU 10-715: Machine Learning (Fall 2015) http://www.cs.cmu.eu/~bapoczos/classes/ml10715_2015fall/ OUT: Oct 5, 2015 DUE: Oct 19, 2015, 10:20 AM Guielines The

More information

Lie symmetry and Mei conservation law of continuum system

Lie symmetry and Mei conservation law of continuum system Chin. Phys. B Vol. 20, No. 2 20 020 Lie symmetry an Mei conservation law of continuum system Shi Shen-Yang an Fu Jing-Li Department of Physics, Zhejiang Sci-Tech University, Hangzhou 3008, China Receive

More information

Separation of Variables

Separation of Variables Physics 342 Lecture 1 Separation of Variables Lecture 1 Physics 342 Quantum Mechanics I Monay, January 25th, 2010 There are three basic mathematical tools we nee, an then we can begin working on the physical

More information

Time-of-Arrival Estimation in Non-Line-Of-Sight Environments

Time-of-Arrival Estimation in Non-Line-Of-Sight Environments 2 Conference on Information Sciences an Systems, The Johns Hopkins University, March 2, 2 Time-of-Arrival Estimation in Non-Line-Of-Sight Environments Sinan Gezici, Hisashi Kobayashi an H. Vincent Poor

More information

SYNCHRONOUS SEQUENTIAL CIRCUITS

SYNCHRONOUS SEQUENTIAL CIRCUITS CHAPTER SYNCHRONOUS SEUENTIAL CIRCUITS Registers an counters, two very common synchronous sequential circuits, are introuce in this chapter. Register is a igital circuit for storing information. Contents

More information

Situation awareness of power system based on static voltage security region

Situation awareness of power system based on static voltage security region The 6th International Conference on Renewable Power Generation (RPG) 19 20 October 2017 Situation awareness of power system base on static voltage security region Fei Xiao, Zi-Qing Jiang, Qian Ai, Ran

More information

Fast Inference and Learning for Modeling Documents with a Deep Boltzmann Machine

Fast Inference and Learning for Modeling Documents with a Deep Boltzmann Machine Fast Inference and Learning for Modeling Documents with a Deep Boltzmann Machine Nitish Srivastava nitish@cs.toronto.edu Ruslan Salahutdinov rsalahu@cs.toronto.edu Geoffrey Hinton hinton@cs.toronto.edu

More information

Optimal Signal Detection for False Track Discrimination

Optimal Signal Detection for False Track Discrimination Optimal Signal Detection for False Track Discrimination Thomas Hanselmann Darko Mušicki Dept. of Electrical an Electronic Eng. Dept. of Electrical an Electronic Eng. The University of Melbourne The University

More information

Latent variable models for discrete data

Latent variable models for discrete data Latent variable models for discrete data Jianfei Chen Department of Computer Science and Technology Tsinghua University, Beijing 100084 chris.jianfei.chen@gmail.com Janurary 13, 2014 Murphy, Kevin P. Machine

More information

Topic Models with Topic Ordering Regularities for Topic Segmentation

Topic Models with Topic Ordering Regularities for Topic Segmentation Topic Moels with Topic Orering Regularities for Topic Segmentation Lan Du Department of computing Macquarie University Syney, NSW 2109, Australia Email: lan.u@mq.eu.au John K Pate Department of Computing

More information

Switching Time Optimization in Discretized Hybrid Dynamical Systems

Switching Time Optimization in Discretized Hybrid Dynamical Systems Switching Time Optimization in Discretize Hybri Dynamical Systems Kathrin Flaßkamp, To Murphey, an Sina Ober-Blöbaum Abstract Switching time optimization (STO) arises in systems that have a finite set

More information

Similarity Measures for Categorical Data A Comparative Study. Technical Report

Similarity Measures for Categorical Data A Comparative Study. Technical Report Similarity Measures for Categorical Data A Comparative Stuy Technical Report Department of Computer Science an Engineering University of Minnesota 4-92 EECS Builing 200 Union Street SE Minneapolis, MN

More information

Collaborative User Clustering for Short Text Streams

Collaborative User Clustering for Short Text Streams Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence (AAAI-17) Collaborative User Clustering for Short Text Streams Shangsong Liang, Zhaochun Ren, Emine Yilmaz, and Evangelos Kanoulas

More information

Lower Bounds for the Smoothed Number of Pareto optimal Solutions

Lower Bounds for the Smoothed Number of Pareto optimal Solutions Lower Bouns for the Smoothe Number of Pareto optimal Solutions Tobias Brunsch an Heiko Röglin Department of Computer Science, University of Bonn, Germany brunsch@cs.uni-bonn.e, heiko@roeglin.org Abstract.

More information

d dx But have you ever seen a derivation of these results? We ll prove the first result below. cos h 1

d dx But have you ever seen a derivation of these results? We ll prove the first result below. cos h 1 Lecture 5 Some ifferentiation rules Trigonometric functions (Relevant section from Stewart, Seventh Eition: Section 3.3) You all know that sin = cos cos = sin. () But have you ever seen a erivation of

More information

Kernel Density Topic Models: Visual Topics Without Visual Words

Kernel Density Topic Models: Visual Topics Without Visual Words Kernel Density Topic Models: Visual Topics Without Visual Words Konstantinos Rematas K.U. Leuven ESAT-iMinds krematas@esat.kuleuven.be Mario Fritz Max Planck Institute for Informatics mfrtiz@mpi-inf.mpg.de

More information

Modeling Real Estate for School District Identification

Modeling Real Estate for School District Identification 2016 IEEE 16th International Conference on Data Mining Moeling Real Estate for School District Ientification Fei Tan, Chaoran Cheng, an Zhi Wei Department of Computer Science, New Jersey Institute of Technology

More information

Optimized Signal De-noising Algorithm for Acoustic Emission Leakage

Optimized Signal De-noising Algorithm for Acoustic Emission Leakage 1009 A publication of CHEMICAL ENGINEERING TRANSACTIONS VOL. 46, 2015 Guest Eitors: Peiyu Ren, Yancang Li, Huiping Song Copyright 2015, AIDIC Servizi S.r.l., ISBN 978-88-95608-37-2; ISSN 2283-9216 The

More information

Study Notes on the Latent Dirichlet Allocation

Study Notes on the Latent Dirichlet Allocation Study Notes on the Latent Dirichlet Allocation Xugang Ye 1. Model Framework A word is an element of dictionary {1,,}. A document is represented by a sequence of words: =(,, ), {1,,}. A corpus is a collection

More information

Nonlinear Adaptive Ship Course Tracking Control Based on Backstepping and Nussbaum Gain

Nonlinear Adaptive Ship Course Tracking Control Based on Backstepping and Nussbaum Gain Nonlinear Aaptive Ship Course Tracking Control Base on Backstepping an Nussbaum Gain Jialu Du, Chen Guo Abstract A nonlinear aaptive controller combining aaptive Backstepping algorithm with Nussbaum gain

More information

A Modification of the Jarque-Bera Test. for Normality

A Modification of the Jarque-Bera Test. for Normality Int. J. Contemp. Math. Sciences, Vol. 8, 01, no. 17, 84-85 HIKARI Lt, www.m-hikari.com http://x.oi.org/10.1988/ijcms.01.9106 A Moification of the Jarque-Bera Test for Normality Moawa El-Fallah Ab El-Salam

More information

Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen. Neural Networks. Tobias Scheffer

Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen. Neural Networks. Tobias Scheffer Universität Potsam Institut für Informatik Lehrstuhl Maschinelles Lernen Neural Networks Tobias Scheffer Overview Neural information processing. Fee-forwar networks. Training fee-forwar networks, back

More information

Necessary and Sufficient Conditions for Sketched Subspace Clustering

Necessary and Sufficient Conditions for Sketched Subspace Clustering Necessary an Sufficient Conitions for Sketche Subspace Clustering Daniel Pimentel-Alarcón, Laura Balzano 2, Robert Nowak University of Wisconsin-Maison, 2 University of Michigan-Ann Arbor Abstract This

More information

Robust Low Rank Kernel Embeddings of Multivariate Distributions

Robust Low Rank Kernel Embeddings of Multivariate Distributions Robust Low Rank Kernel Embeings of Multivariate Distributions Le Song, Bo Dai College of Computing, Georgia Institute of Technology lsong@cc.gatech.eu, boai@gatech.eu Abstract Kernel embeing of istributions

More information

Optimized Schwarz Methods with the Yin-Yang Grid for Shallow Water Equations

Optimized Schwarz Methods with the Yin-Yang Grid for Shallow Water Equations Optimize Schwarz Methos with the Yin-Yang Gri for Shallow Water Equations Abessama Qaouri Recherche en prévision numérique, Atmospheric Science an Technology Directorate, Environment Canaa, Dorval, Québec,

More information

A Unified Posterior Regularized Topic Model with Maximum Margin for Learning-to-Rank

A Unified Posterior Regularized Topic Model with Maximum Margin for Learning-to-Rank A Unified Posterior Regularized Topic Model with Maximum Margin for Learning-to-Rank Shoaib Jameel Shoaib Jameel 1, Wai Lam 2, Steven Schockaert 1, and Lidong Bing 3 1 School of Computer Science and Informatics,

More information

IPSJ SIG Technical Report Vol.2014-MPS-100 No /9/25 1,a) 1 1 SNS / / / / / / Time Series Topic Model Considering Dependence to Multiple Topics S

IPSJ SIG Technical Report Vol.2014-MPS-100 No /9/25 1,a) 1 1 SNS / / / / / / Time Series Topic Model Considering Dependence to Multiple Topics S 1,a) 1 1 SNS /// / // Time Series Topic Model Considering Dependence to Multiple Topics Sasaki Kentaro 1,a) Yoshikawa Tomohiro 1 Furuhashi Takeshi 1 Abstract: This pater proposes a topic model that considers

More information

LATTICE-BASED D-OPTIMUM DESIGN FOR FOURIER REGRESSION

LATTICE-BASED D-OPTIMUM DESIGN FOR FOURIER REGRESSION The Annals of Statistics 1997, Vol. 25, No. 6, 2313 2327 LATTICE-BASED D-OPTIMUM DESIGN FOR FOURIER REGRESSION By Eva Riccomagno, 1 Rainer Schwabe 2 an Henry P. Wynn 1 University of Warwick, Technische

More information

Designing of Acceptance Double Sampling Plan for Life Test Based on Percentiles of Exponentiated Rayleigh Distribution

Designing of Acceptance Double Sampling Plan for Life Test Based on Percentiles of Exponentiated Rayleigh Distribution International Journal of Statistics an Systems ISSN 973-675 Volume, Number 3 (7), pp. 475-484 Research Inia Publications http://www.ripublication.com Designing of Acceptance Double Sampling Plan for Life

More information

Equilibrium in Queues Under Unknown Service Times and Service Value

Equilibrium in Queues Under Unknown Service Times and Service Value University of Pennsylvania ScholarlyCommons Finance Papers Wharton Faculty Research 1-2014 Equilibrium in Queues Uner Unknown Service Times an Service Value Laurens Debo Senthil K. Veeraraghavan University

More information

Construction of the Electronic Radial Wave Functions and Probability Distributions of Hydrogen-like Systems

Construction of the Electronic Radial Wave Functions and Probability Distributions of Hydrogen-like Systems Construction of the Electronic Raial Wave Functions an Probability Distributions of Hyrogen-like Systems Thomas S. Kuntzleman, Department of Chemistry Spring Arbor University, Spring Arbor MI 498 tkuntzle@arbor.eu

More information

Lectures - Week 10 Introduction to Ordinary Differential Equations (ODES) First Order Linear ODEs

Lectures - Week 10 Introduction to Ordinary Differential Equations (ODES) First Order Linear ODEs Lectures - Week 10 Introuction to Orinary Differential Equations (ODES) First Orer Linear ODEs When stuying ODEs we are consiering functions of one inepenent variable, e.g., f(x), where x is the inepenent

More information

Hybrid Fusion for Biometrics: Combining Score-level and Decision-level Fusion

Hybrid Fusion for Biometrics: Combining Score-level and Decision-level Fusion Hybri Fusion for Biometrics: Combining Score-level an Decision-level Fusion Qian Tao Raymon Velhuis Signals an Systems Group, University of Twente Postbus 217, 7500AE Enschee, the Netherlans {q.tao,r.n.j.velhuis}@ewi.utwente.nl

More information

The derivative of a function f(x) is another function, defined in terms of a limiting expression: f(x + δx) f(x)

The derivative of a function f(x) is another function, defined in terms of a limiting expression: f(x + δx) f(x) Y. D. Chong (2016) MH2801: Complex Methos for the Sciences 1. Derivatives The erivative of a function f(x) is another function, efine in terms of a limiting expression: f (x) f (x) lim x δx 0 f(x + δx)

More information

A Unified Approach for Learning the Parameters of Sum-Product Networks

A Unified Approach for Learning the Parameters of Sum-Product Networks A Unifie Approach for Learning the Parameters of Sum-Prouct Networks Han Zhao Machine Learning Dept. Carnegie Mellon University han.zhao@cs.cmu.eu Pascal Poupart School of Computer Science University of

More information

CONTROL CHARTS FOR VARIABLES

CONTROL CHARTS FOR VARIABLES UNIT CONTOL CHATS FO VAIABLES Structure.1 Introuction Objectives. Control Chart Technique.3 Control Charts for Variables.4 Control Chart for Mean(-Chart).5 ange Chart (-Chart).6 Stanar Deviation Chart

More information

A Note on Exact Solutions to Linear Differential Equations by the Matrix Exponential

A Note on Exact Solutions to Linear Differential Equations by the Matrix Exponential Avances in Applie Mathematics an Mechanics Av. Appl. Math. Mech. Vol. 1 No. 4 pp. 573-580 DOI: 10.4208/aamm.09-m0946 August 2009 A Note on Exact Solutions to Linear Differential Equations by the Matrix

More information

Lecture Introduction. 2 Examples of Measure Concentration. 3 The Johnson-Lindenstrauss Lemma. CS-621 Theory Gems November 28, 2012

Lecture Introduction. 2 Examples of Measure Concentration. 3 The Johnson-Lindenstrauss Lemma. CS-621 Theory Gems November 28, 2012 CS-6 Theory Gems November 8, 0 Lecture Lecturer: Alesaner Mąry Scribes: Alhussein Fawzi, Dorina Thanou Introuction Toay, we will briefly iscuss an important technique in probability theory measure concentration

More information

NOTES ON EULER-BOOLE SUMMATION (1) f (l 1) (n) f (l 1) (m) + ( 1)k 1 k! B k (y) f (k) (y) dy,

NOTES ON EULER-BOOLE SUMMATION (1) f (l 1) (n) f (l 1) (m) + ( 1)k 1 k! B k (y) f (k) (y) dy, NOTES ON EULER-BOOLE SUMMATION JONATHAN M BORWEIN, NEIL J CALKIN, AND DANTE MANNA Abstract We stuy a connection between Euler-MacLaurin Summation an Boole Summation suggeste in an AMM note from 196, which

More information

Thermal conductivity of graded composites: Numerical simulations and an effective medium approximation

Thermal conductivity of graded composites: Numerical simulations and an effective medium approximation JOURNAL OF MATERIALS SCIENCE 34 (999)5497 5503 Thermal conuctivity of grae composites: Numerical simulations an an effective meium approximation P. M. HUI Department of Physics, The Chinese University

More information

Fractional Geometric Calculus: Toward A Unified Mathematical Language for Physics and Engineering

Fractional Geometric Calculus: Toward A Unified Mathematical Language for Physics and Engineering Fractional Geometric Calculus: Towar A Unifie Mathematical Language for Physics an Engineering Xiong Wang Center of Chaos an Complex Network, Department of Electronic Engineering, City University of Hong

More information

Lower bounds on Locality Sensitive Hashing

Lower bounds on Locality Sensitive Hashing Lower bouns on Locality Sensitive Hashing Rajeev Motwani Assaf Naor Rina Panigrahy Abstract Given a metric space (X, X ), c 1, r > 0, an p, q [0, 1], a istribution over mappings H : X N is calle a (r,

More information

Proof of SPNs as Mixture of Trees

Proof of SPNs as Mixture of Trees A Proof of SPNs as Mixture of Trees Theorem 1. If T is an inuce SPN from a complete an ecomposable SPN S, then T is a tree that is complete an ecomposable. Proof. Argue by contraiction that T is not a

More information

Axiometrics: Axioms of Information Retrieval Effectiveness Metrics

Axiometrics: Axioms of Information Retrieval Effectiveness Metrics Axiometrics: Axioms of Information Retrieval Effectiveness Metrics ABSTRACT Ey Maalena Department of Maths Computer Science University of Uine Uine, Italy ey.maalena@uniu.it There are literally ozens most

More information

Text Mining for Economics and Finance Latent Dirichlet Allocation

Text Mining for Economics and Finance Latent Dirichlet Allocation Text Mining for Economics and Finance Latent Dirichlet Allocation Stephen Hansen Text Mining Lecture 5 1 / 45 Introduction Recall we are interested in mixed-membership modeling, but that the plsi model

More information

Cross-domain Collaboration Recommendation

Cross-domain Collaboration Recommendation Cross- Collaboration Recommenation Jie Tang,SenWu, Jimeng Sun, an Hang Su Department of Computer Science an Technology, Tsinghua University IBM TJ Watson Research Center, USA jietang@tsinghua.eu.cn, ronalosen@gmail.com,

More information

Classifying Biomedical Text Abstracts based on Hierarchical Concept Structure

Classifying Biomedical Text Abstracts based on Hierarchical Concept Structure Classifying Biomeical Text Abstracts base on Hierarchical Concept Structure Rozilawati Binti Dollah an Masai Aono Abstract Classifying biomeical literature is a ifficult an challenging tas, especially

More information

Attribute Reduction and Information Granularity *

Attribute Reduction and Information Granularity * ttribute Reuction an nformation Granularity * Li-hong Wang School of Computer Engineering an Science, Shanghai University, Shanghai, 200072, P.R.C School of Computer Science an Technology, Yantai University,

More information

Lecture 6: Calculus. In Song Kim. September 7, 2011

Lecture 6: Calculus. In Song Kim. September 7, 2011 Lecture 6: Calculus In Song Kim September 7, 20 Introuction to Differential Calculus In our previous lecture we came up with several ways to analyze functions. We saw previously that the slope of a linear

More information

Calculus of Variations

Calculus of Variations 16.323 Lecture 5 Calculus of Variations Calculus of Variations Most books cover this material well, but Kirk Chapter 4 oes a particularly nice job. x(t) x* x*+ αδx (1) x*- αδx (1) αδx (1) αδx (1) t f t

More information

A simple model for the small-strain behaviour of soils

A simple model for the small-strain behaviour of soils A simple moel for the small-strain behaviour of soils José Jorge Naer Department of Structural an Geotechnical ngineering, Polytechnic School, University of São Paulo 05508-900, São Paulo, Brazil, e-mail:

More information

Learning the Semantic Correlation: An Alternative Way to Gain from Unlabeled Text

Learning the Semantic Correlation: An Alternative Way to Gain from Unlabeled Text Learning the Semantic Correlation: An Alternative Way to Gain from Unlabeled Text Yi Zhang Machine Learning Department Carnegie Mellon University yizhang1@cs.cmu.edu Jeff Schneider The Robotics Institute

More information

Latent Dirichlet Allocation Introduction/Overview

Latent Dirichlet Allocation Introduction/Overview Latent Dirichlet Allocation Introduction/Overview David Meyer 03.10.2016 David Meyer http://www.1-4-5.net/~dmm/ml/lda_intro.pdf 03.10.2016 Agenda What is Topic Modeling? Parametric vs. Non-Parametric Models

More information

arxiv:physics/ v2 [physics.ed-ph] 23 Sep 2003

arxiv:physics/ v2 [physics.ed-ph] 23 Sep 2003 Mass reistribution in variable mass systems Célia A. e Sousa an Vítor H. Rorigues Departamento e Física a Universiae e Coimbra, P-3004-516 Coimbra, Portugal arxiv:physics/0211075v2 [physics.e-ph] 23 Sep

More information

The influence of the equivalent hydraulic diameter on the pressure drop prediction of annular test section

The influence of the equivalent hydraulic diameter on the pressure drop prediction of annular test section IOP Conference Series: Materials Science an Engineering PAPER OPEN ACCESS The influence of the equivalent hyraulic iameter on the pressure rop preiction of annular test section To cite this article: A

More information

2Algebraic ONLINE PAGE PROOFS. foundations

2Algebraic ONLINE PAGE PROOFS. foundations Algebraic founations. Kick off with CAS. Algebraic skills.3 Pascal s triangle an binomial expansions.4 The binomial theorem.5 Sets of real numbers.6 Surs.7 Review . Kick off with CAS Playing lotto Using

More information

Topic Modelling and Latent Dirichlet Allocation

Topic Modelling and Latent Dirichlet Allocation Topic Modelling and Latent Dirichlet Allocation Stephen Clark (with thanks to Mark Gales for some of the slides) Lent 2013 Machine Learning for Language Processing: Lecture 7 MPhil in Advanced Computer

More information

FLUCTUATIONS IN THE NUMBER OF POINTS ON SMOOTH PLANE CURVES OVER FINITE FIELDS. 1. Introduction

FLUCTUATIONS IN THE NUMBER OF POINTS ON SMOOTH PLANE CURVES OVER FINITE FIELDS. 1. Introduction FLUCTUATIONS IN THE NUMBER OF POINTS ON SMOOTH PLANE CURVES OVER FINITE FIELDS ALINA BUCUR, CHANTAL DAVID, BROOKE FEIGON, MATILDE LALÍN 1 Introuction In this note, we stuy the fluctuations in the number

More information

Gaussian processes with monotonicity information

Gaussian processes with monotonicity information Gaussian processes with monotonicity information Anonymous Author Anonymous Author Unknown Institution Unknown Institution Abstract A metho for using monotonicity information in multivariate Gaussian process

More information

Generalizing Kronecker Graphs in order to Model Searchable Networks

Generalizing Kronecker Graphs in order to Model Searchable Networks Generalizing Kronecker Graphs in orer to Moel Searchable Networks Elizabeth Boine, Babak Hassibi, Aam Wierman California Institute of Technology Pasaena, CA 925 Email: {eaboine, hassibi, aamw}@caltecheu

More information

arxiv: v4 [cs.ds] 7 Mar 2014

arxiv: v4 [cs.ds] 7 Mar 2014 Analysis of Agglomerative Clustering Marcel R. Ackermann Johannes Blömer Daniel Kuntze Christian Sohler arxiv:101.697v [cs.ds] 7 Mar 01 Abstract The iameter k-clustering problem is the problem of partitioning

More information

A Simple Model for the Calculation of Plasma Impedance in Atmospheric Radio Frequency Discharges

A Simple Model for the Calculation of Plasma Impedance in Atmospheric Radio Frequency Discharges Plasma Science an Technology, Vol.16, No.1, Oct. 214 A Simple Moel for the Calculation of Plasma Impeance in Atmospheric Raio Frequency Discharges GE Lei ( ) an ZHANG Yuantao ( ) Shanong Provincial Key

More information

Math 1B, lecture 8: Integration by parts

Math 1B, lecture 8: Integration by parts Math B, lecture 8: Integration by parts Nathan Pflueger 23 September 2 Introuction Integration by parts, similarly to integration by substitution, reverses a well-known technique of ifferentiation an explores

More information