Learning topic models by neighborhood aggregation

3Citations
Citations of this article
30Readers
Mendeley users who have this article in their library.

Abstract

Topic models are frequently used in machine learning owing to their high interpretability and modular structure. However, extending a topic model to include a supervisory signal, to incorporate pre-trained word embedding vectors and to include a nonlinear output function is not an easy task because one has to resort to a highly intricate approximate inference procedure. The present paper shows that topic modeling with pre-trained word embedding vectors can be viewed as implementing a neighborhood aggregation algorithm where messages are passed through a network defined over words. From the network view of topic models, nodes correspond to words in a document and edges correspond to either a relationship describing co-occurring words in a document or a relationship describing the same word in the corpus. The network view allows us to extend the model to include supervisory signals, incorporate pre-trained word embedding vectors and include a nonlinear output function in a simple manner. In experiments, we show that our approach outperforms the state-of-the-art supervised Latent Dirichlet Allocation implementation in terms of held-out document classification tasks.

Cite

CITATION STYLE

APA

Hisano, R. (2019). Learning topic models by neighborhood aggregation. In IJCAI International Joint Conference on Artificial Intelligence (Vol. 2019-August, pp. 2498–2505). International Joint Conferences on Artificial Intelligence. https://doi.org/10.24963/ijcai.2019/347

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free