SAE-NTM: Sentence-Aware Encoder for Neural Topic Modeling

4Citations
Citations of this article
12Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Incorporating external knowledge, such as pre-trained language models (PLMs), into neural topic modeling has achieved great success in recent years. However, employing PLMs for topic modeling generally ignores the maximum sequence length of PLMs and the interaction between external knowledge and bag-of-words (BOW). To this end, we propose a sentence-aware encoder for neural topic modeling, which adopts fine-grained sentence embeddings as external knowledge to entirely utilize the semantic information of input documents. We introduce sentence-aware attention for document representation, where BOW enables the model to attend on topical sentences that convey topic-related cues. Experiments on three benchmark datasets show that our framework outperforms other state-of-the-art neural topic models in topic coherence. Further, we demonstrate that the proposed approach can yield better latent document-topic features through improvement on the document classification.

Cite

CITATION STYLE

APA

Liu, H., Gao, J., Xiang, S., Liu, T., & Fu, Y. (2023). SAE-NTM: Sentence-Aware Encoder for Neural Topic Modeling. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 106–111). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.codi-1.14

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free