2017
Estimating Word probabilities with neural networks in latent dirichlet allocation
Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
- Volume
- 10526
- Number
- First page
- 129
- Last page
- 137
- Language
- English
- Publishing type
- Research paper (international conference proceedings)
- DOI
- 10.1007/978-3-319-67274-8_12
- Publisher
- Springer Verlag
This paper proposes a new method for estimating the word probabilities in latent Dirichlet allocation (LDA). LDA uses a Dirichlet distribution as the prior for the per-document topic discrete distributions. While another Dirichlet prior can be introduced for the per-topic word discrete distributions, point estimations may lead to a better evaluation result, e.g. in terms of test perplexity. This paper proposes a method for the point estimation of the per-topic word probabilities in LDA by using multilayer perceptron (MLP). Our point estimation is performed in an online manner by mini-batch gradient ascent. We compared our method to the baseline method using a perceptron with no hidden layers and also to the collapsed Gibbs sampling (CGS). The evaluation experiment showed that the test perplexity of CGS could not be improved in almost all cases. However, there certainly were situations where our method achieved a better perplexity than the baseline. We also discuss a usage of our method as word embedding.
- Link information
- ID information
-
- DOI : 10.1007/978-3-319-67274-8_12
- ISSN : 1611-3349
- ISSN : 0302-9743
- DBLP ID : conf/pakdd/Masada17
- SCOPUS ID : 85031430169