Papers

Peer-reviewed
2017

Estimating Word probabilities with neural networks in latent dirichlet allocation

Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
  • Tomonari Masada

Volume
10526
Number
First page
129
Last page
137
Language
English
Publishing type
Research paper (international conference proceedings)
DOI
10.1007/978-3-319-67274-8_12
Publisher
Springer Verlag

This paper proposes a new method for estimating the word probabilities in latent Dirichlet allocation (LDA). LDA uses a Dirichlet distribution as the prior for the per-document topic discrete distributions. While another Dirichlet prior can be introduced for the per-topic word discrete distributions, point estimations may lead to a better evaluation result, e.g. in terms of test perplexity. This paper proposes a method for the point estimation of the per-topic word probabilities in LDA by using multilayer perceptron (MLP). Our point estimation is performed in an online manner by mini-batch gradient ascent. We compared our method to the baseline method using a perceptron with no hidden layers and also to the collapsed Gibbs sampling (CGS). The evaluation experiment showed that the test perplexity of CGS could not be improved in almost all cases. However, there certainly were situations where our method achieved a better perplexity than the baseline. We also discuss a usage of our method as word embedding.

Link information
DOI
https://doi.org/10.1007/978-3-319-67274-8_12
DBLP
https://dblp.uni-trier.de/rec/conf/pakdd/Masada17
URL
http://dblp.uni-trier.de/db/conf/pakdd/pakdd2017-w.html#conf/pakdd/Masada17
ID information
  • DOI : 10.1007/978-3-319-67274-8_12
  • ISSN : 1611-3349
  • ISSN : 0302-9743
  • DBLP ID : conf/pakdd/Masada17
  • SCOPUS ID : 85031430169

Export
BibTeX RIS