Applying hlda to Practical Topic Modeling

Similar documents
Latent Dirichlet Allocation (LDA)

Topic Modelling and Latent Dirichlet Allocation

Applying LDA topic model to a corpus of Italian Supreme Court decisions

Non-Parametric Bayes

Content-based Recommendation

LDA with Amortized Inference

Latent Dirichlet Allocation (LDA)

Dirichlet Enhanced Latent Semantic Analysis

CS Lecture 18. Topic Models and LDA

Lecture 19, November 19, 2012

Text Mining for Economics and Finance Latent Dirichlet Allocation

Latent Dirichlet Allocation Introduction/Overview

Sparse Stochastic Inference for Latent Dirichlet Allocation

Study Notes on the Latent Dirichlet Allocation

16 : Approximate Inference: Markov Chain Monte Carlo

Non-parametric Clustering with Dirichlet Processes

Decoupling Sparsity and Smoothness in the Discrete Hierarchical Dirichlet Process

Evaluation Methods for Topic Models

Topic Models and Applications to Short Documents

Information retrieval LSI, plsi and LDA. Jian-Yun Nie

Bayesian Nonparametrics for Speech and Signal Processing

arxiv: v1 [stat.ml] 8 Jan 2012

Advanced Machine Learning

Lecture 13 : Variational Inference: Mean Field Approximation

Nested Hierarchical Dirichlet Processes

Gaussian Mixture Model

Latent Dirichlet Allocation

28 : Approximate Inference - Distributed MCMC

Topic Modeling Using Latent Dirichlet Allocation (LDA)

Distance dependent Chinese restaurant processes

Bayesian Nonparametric Models

COMS 4721: Machine Learning for Data Science Lecture 18, 4/4/2017

arxiv: v2 [stat.ml] 5 Nov 2012

Applying Latent Dirichlet Allocation to Group Discovery in Large Graphs

Collapsed Variational Inference for HDP

Document and Topic Models: plsa and LDA

Scalable Training of Hierarchical Topic Models

LDA Based Similarity Modeling for Question Answering

Pachinko Allocation: DAG-Structured Mixture Models of Topic Correlations

Latent Dirichlet Alloca/on

Machine Learning for Data Science (CS4786) Lecture 24

GLAD: Group Anomaly Detection in Social Media Analysis

AN INTRODUCTION TO TOPIC MODELS

Replicated Softmax: an Undirected Topic Model. Stephen Turner

Topic Models. Advanced Machine Learning for NLP Jordan Boyd-Graber OVERVIEW. Advanced Machine Learning for NLP Boyd-Graber Topic Models 1 of 1

Efficient Tree-Based Topic Modeling

Fast Inference and Learning for Modeling Documents with a Deep Boltzmann Machine

Topic Models. Charles Elkan November 20, 2008

13: Variational inference II

Bayesian Nonparametrics: Dirichlet Process

A Unified Posterior Regularized Topic Model with Maximum Margin for Learning-to-Rank

Chapter 8 PROBABILISTIC MODELS FOR TEXT MINING. Yizhou Sun Department of Computer Science University of Illinois at Urbana-Champaign

Latent Dirichlet Allocation Based Multi-Document Summarization

Distributed ML for DOSNs: giving power back to users

Topic Learning and Inference Using Dirichlet Allocation Product Partition Models and Hybrid Metropolis Search

Probabilistic Latent Semantic Analysis

PROBABILISTIC LATENT SEMANTIC ANALYSIS

Probabilistic Topic Models Tutorial: COMAD 2011

Statistical Debugging with Latent Topic Models

Lecture 3a: Dirichlet processes

MCMC and Gibbs Sampling. Kayhan Batmanghelich

Hierarchical Dirichlet Processes

Crouching Dirichlet, Hidden Markov Model: Unsupervised POS Tagging with Context Local Tag Generation

Generative Clustering, Topic Modeling, & Bayesian Inference

Deep Variational Inference. FLARE Reading Group Presentation Wesley Tansey 9/28/2016

Introduction to Bayesian inference

Chapter 4 Dynamic Bayesian Networks Fall Jin Gu, Michael Zhang

Gibbs Sampling. Héctor Corrada Bravo. University of Maryland, College Park, USA CMSC 644:

Measuring Topic Quality in Latent Dirichlet Allocation

Hierarchical Models, Nested Models and Completely Random Measures

Online but Accurate Inference for Latent Variable Models with Local Gibbs Sampling

Collaborative Topic Modeling for Recommending Scientific Articles

A Continuous-Time Model of Topic Co-occurrence Trends

Infering the Number of State Clusters in Hidden Markov Model and its Extension

27 : Distributed Monte Carlo Markov Chain. 1 Recap of MCMC and Naive Parallel Gibbs Sampling

Latent Dirichlet Bayesian Co-Clustering

Mixtures of Multinomials

Learning Energy-Based Models of High-Dimensional Data

Topic Models. Brandon Malone. February 20, Latent Dirichlet Allocation Success Stories Wrap-up

Latent Dirichlet Allocation

Image segmentation combining Markov Random Fields and Dirichlet Processes

Linear Dynamical Systems

Nonparametric Bayes Pachinko Allocation

STA 4273H: Statistical Machine Learning

Construction of Dependent Dirichlet Processes based on Poisson Processes

Collapsed Variational Bayesian Inference for Hidden Markov Models

CS 188: Artificial Intelligence. Bayes Nets

Online Bayesian Passive-Agressive Learning

Optimization Number of Topic Latent Dirichlet Allocation

Topic Models. Material adapted from David Mimno University of Maryland INTRODUCTION. Material adapted from David Mimno UMD Topic Models 1 / 51

Lecturer: David Blei Lecture #3 Scribes: Jordan Boyd-Graber and Francisco Pereira October 1, 2007

CSCI 5822 Probabilistic Model of Human and Machine Learning. Mike Mozer University of Colorado

Kernel Density Topic Models: Visual Topics Without Visual Words

Streaming Variational Bayes

Graphical Models and Kernel Methods

Graphical Models for Query-driven Analysis of Multimodal Data

Introduction to Machine Learning

Bayesian Learning and Inference in Recurrent Switching Linear Dynamical Systems

Clustering problems, mixture models and Bayesian nonparametrics

Recent Advances in Bayesian Inference Techniques

Transcription:

Joseph Heng lengerfulluse@gmail.com CIST Lab of BUPT March 17, 2013

Outline 1 HLDA Discussion 2 the nested CRP GEM Distribution Dirichlet Distribution Posterior Inference

Outline 1 HLDA Discussion 2 the nested CRP GEM Distribution Dirichlet Distribution Posterior Inference 3 Manual Modeling Procedure Empirical Results

Outline 1 HLDA Discussion 2 the nested CRP GEM Distribution Dirichlet Distribution Posterior Inference 3 Manual Modeling Procedure Empirical Results 4

Outline 1 HLDA Discussion 2 the nested CRP GEM Distribution Dirichlet Distribution Posterior Inference 3 Manual Modeling Procedure Empirical Results 4

Background HLDA Discussion The Goal in Our Paper. HLDA has been proved to be a powerful tool. One of the bottlenecks which prevent its large-scale application is that we cannot find a quick and effective approach to modeling new data properly. There exist lots of factors, eg. hyper-parameter settings, uncertainty of random algorithms and features of different corpus. Probabilistic Topic Models Topic models are algorithms for discovering the main themes that pervade a large and otherwise unstructured collection of documents. Topic models can organize the collection according to the discovered themes. LDA-related LDA and other topic models are part of the larger field of probabilistic modeling. In generative probabilistic modeling, we treat our data as arising from a generative process that includes hidden variables.

Merits of HLDA HLDA Discussion 1 Generative process for documents [2] 2 Posterior approximate inference with Gibbs sampling [1] p(z d,n z (d,n), c, w, π, η) p(z d,n z d, n, m, π)p(z d,n z, c, w (d,n), η) (1) p(c d w, c d, z, η, γ) p(c d c d, γ)p(w d c, w d, z, η) (2)

Merits of HLDA HLDA Discussion 1 Generative process for documents [2] 2 Posterior approximate inference with Gibbs sampling [1] p(z d,n z (d,n), c, w, π, η) p(z d,n z d, n, m, π)p(z d,n z, c, w (d,n), η) (1) p(c d w, c d, z, η, γ) p(c d c d, γ)p(w d c, w d, z, η) (2) 3 Assessing convergence and approximating the mode. L (t) = log p(c (t) 1:D, z(t) 1:D, z 1:D η, γ, m, π) (3)

Merits of HLDA HLDA Discussion 1 Generative process for documents [2] 2 Posterior approximate inference with Gibbs sampling [1] p(z d,n z (d,n), c, w, π, η) p(z d,n z d, n, m, π)p(z d,n z, c, w (d,n), η) (1) p(c d w, c d, z, η, γ) p(c d c d, γ)p(w d c, w d, z, η) (2) 3 Assessing convergence and approximating the mode. L (t) = log p(c (t) 1:D, z(t) 1:D, z 1:D η, γ, m, π) (3)

Practical Difficulty HLDA Discussion What s practical problem [3] when using hdla to topic modeling? Why unified framework?

Practical Difficulty HLDA Discussion What s practical problem [3] when using hdla to topic modeling? Why unified framework?

Practical Difficulty HLDA Discussion What s practical problem [3] when using hdla to topic modeling? Why unified framework? Figure : unified analysis framework with two clues

Generative Process the nested CRP GEM Distribution Dirichlet Distribution Posterior Inference Document are assumed to be drawn from the following process. For each node k T in the infinite tree, draw a topic β k Dirichlet(γ). For each document, d 1,2,...,D Draw C d ncrp(γ) to choose the path Draw a distribution over levels in the tree, θ d m, π GEM(m, π). For each word, Choose level Z d,n θ Mult(θ d ). Choose word W d,n Z d,n, C d, β Mult(β cd, [Z d,n ).

ncrp the nested CRP GEM Distribution Dirichlet Distribution Posterior Inference to CRP algorithm γ parameters Experiments with γ Comparision

GEM the nested CRP GEM Distribution Dirichlet Distribution Posterior Inference The Different View of DP Parameters m and π Experiments with m and π

Dirichlet Process the nested CRP GEM Distribution Dirichlet Distribution Posterior Inference 1 Experiment with parameter 2 Relationship with Three above.

Iterator Convergency the nested CRP GEM Distribution Dirichlet Distribution Posterior Inference Monte Carlo Markov Chain (Collapsed) Gibbs Sampling Algorithm

Manual Modeling Procedure Empirical Results Tree Depth

Manual Modeling Procedure Empirical Results Tree Depth theme\depth 3 4 5 XTUS 4.9 59.7 78.7 IRSA 3.6 35.8 46.5 LTFC 7.1 54.2 74.8 EUDC 4.1 59.2 83.5 GBAB 3.1 47 66.2

Manual Modeling Procedure Empirical Results Tree Depth Sampling or not

Manual Modeling Procedure Empirical Results Tree Depth Sampling or not

Manual Modeling Procedure Empirical Results Tree Depth Sampling or not Corpus features

Manual Modeling Procedure Empirical Results Tree Depth Sampling or not Corpus features

Manual Modeling Procedure Manual Modeling Procedure Empirical Results Generate hlda input and extract features from corpus. Approximate depth of tree. Topic parameter for each level. ncrp parameter for non-leaf levels. m, π parameter for words allocations. Sampling for hyper-parameters or not.

Empirical Results Manual Modeling Procedure Empirical Results Experiments have been conducted with the guide of modeling procedure above with only three modifications to the settings. theme level#1 level#2 hlda#1 hlda#2 score XTUS 4 9 5.3 10 4 EUDC 4 5 5.2 9 3 IRSA 2 5 3.9 8 4 HCDH 5 7 4.8 10 4 CQWF 4 9 5.3 10 5 SBAG 4 10 6 10 5 GBAB 6 8 4.8 11 3 LTFC 6 14 7.2 12 5 MACO 4 7 6.7 8 4 NOHN 5 9 7 9 4

Asli C and Dilek H. A hybrid hierarchical model for multi-document summarization. ACL 10 Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics, pages 815 824, 2010. Blei David, Andrew Ng, and Jordan. Latent dirichlet allocation. Journal of Machine Learning Research, 2003. Paisley J, Wang C, Blei D M, and Jordan. Nested hierarchical dirichlet proceses. arxiv preprint arxiv:1210.6738, 2012.