Download presentation
Presentation is loading. Please wait.
Published byAnnabella Tucker Modified over 9 years ago
1
Decoupling Sparsity and Smoothness in the Discrete Hierarchical Dirichlet Process Chong Wang and David M. Blei NIPS 2009 Discussion led by Chunping Wang ECE, Duke University March 26, 2010
2
Outline Motivations LDA and HDP-LDA Sparse Topic Models Inference Using Collapsed Gibbs sampling Experiments Conclusions 1/16
3
Motivations 2/16 Topics modeling with the “bag of words” assumption An extension of the HDP-LDA model In the LDA and the HDP-LDA models, the topics are drawn from an exchangeable Dirichlet distribution with a scale parameter. As approaches zero, topics will be o sparse: most probability mass on only a few terms o less smooth: empirical counts dominant Goal: to decouple sparsity and smoothness so that these two properties can be achieved at the same time. How: a Bernoulli variable for each term and each topic is introduced.
4
LDA and HDP-LDA 3/16 LDA HDP-LDA topic : document : word : topic : document : word : Nonparametric form of LDA, with the number of topics unbounded Base measure weights
5
Sparse Topic Models 4/16 The size of the vocabulary is V Defined on a V-1-simplexDefined on a sub-simplex specified by : a V-length binary vector composed of V Bernoulli variables one selection proportion for each topic Sparsity: the pattern of ones in, controlled by Smoothness: enforced over terms with non-zero ’s through Decoupled!
6
Sparse Topic Models 5/16
7
Inference Using Collapsed Gibbs sampling 6/16
8
Inference Using Collapsed Gibbs sampling 6/16 As in the HDP-LDA Topic proportions and topic distributions are integrated out.
9
Inference Using Collapsed Gibbs sampling 6/16 Topic proportions and topic distributions are integrated out. The direct-assignment method based on the Chinese restaurant franchise (CRF) is used for and an augmented variable, table counts As in the HDP-LDA
10
Inference Using Collapsed Gibbs sampling 7/16 Notation: : # of customers (words) in restaurant d (document) eating dish k (topic) : # of tables in restaurant d serving dish k : marginal counts represented with dots K, u: current # of topics and new topic index, respectively : # of times that term v has been assigned to topic k : # of times that all the terms have been assigned to topic k conditional density of under the topic k given all data except
11
Inference Using Collapsed Gibbs sampling 8/16 Recall the direct-assignment sampling method for the HDP-LDA Sampling topic assignments if a new topic is sampled, then sample, and let and and Sampling stick length Sampling table counts
12
Inference Using Collapsed Gibbs sampling 8/16 Recall the direct-assignment sampling method for HDP-LDA Sampling topic assignments for HDP-LDA for sparse TM Instead, the authors integrate out for faster convergence. Since there are total possible, this is the central computational challenge for the sparse TM. straightforward
13
Inference Using Collapsed Gibbs sampling 9/16 where define vocabulary set of terms that have word assignments in topic k This conditional probability depends on the selector proportions.
14
Inference Using Collapsed Gibbs sampling 10/16
15
Inference Using Collapsed Gibbs sampling 10/16
16
Inference Using Collapsed Gibbs sampling 11/16 Sampling Bernoulli parameter ( using as an auxiliary variable) Sampling hyper-parameters o : with Gamma(1,1) priors o : Metropolis-Hastings using symmetric Gaussian proposal Estimate topic distributions from any single sample of z and b define set of terms with an “on” b o sample conditioned on ; o sample conditioned on. sparsity smoothness on the selected terms
17
Experiments 12/16 arXiv: online research abstracts, D = 2500, V = 2873 Nematode Biology: research abstracts, D = 2500, V = 2944 NIPS: NIPS articles between 1988-1999, V = 5005. 20% of words for each paper are used. Conf. abstracts: abstracts from CIKM, ICML, KDD, NIPS, SIGIR and WWW, between 2005-2008, V = 3733. Four datasets: Two predictive quantities: where the topic complexity
18
Experiments 13/16 better perplexity, simpler models larger : smoother less topics similar # of terms
19
Experiments 14/16
20
Experiments 15/16 small (<0.01)
21
Experiments 15/16 small (<0.01) lack of smoothness
22
Experiments 15/16 small (<0.01) Need more topics to explain all kinds of patterns of empirical word counts lack of smoothness
23
Experiments 15/16 Infrequent words populate “noise” topics. small (<0.01) Need more topics to explain all kinds of patterns of empirical word counts lack of smoothness
24
Conclusions 16/16 A new topic model in the HDP-LDA framework, based on the “bag of words” assumption; Main contributions: Decoupling the control of sparsity and smoothness by introducing binary selectors for term assignments in each topic; Developing a collapsed Gibbs sampler in the HDP- LDA framework. Held out performance is better than the HDP-LDA.
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.