Supervised Clustering Loss for Clustering-Friendly Sentence Embeddings: an Application to Intent Clustering

1Citations
Citations of this article
12Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Modern virtual assistants are trained to classify customer requests into a taxonomy of predesigned intents. Requests that fall outside of this taxonomy, however, are often unhandled and need to be clustered to define new experiences. Recently, state-of-the-art results in intent clustering were achieved by training a neural network with a latent structured prediction loss. Unfortunately, though, this new approach suffers from a quadratic bottleneck as it requires to compute a joint embedding representation for all pairs of utterances to cluster. To overcome this limitation, we instead cast the problem into a representation learning task, and we adapt the latent structured prediction loss to fine-tune sentence encoders, thus making it possible to obtain clustering-friendly single-sentence embeddings. Our experiments show that the supervised clustering loss returns state-of-the-art results in terms of clustering accuracy and adjusted mutual information.

Cite

CITATION STYLE

APA

Barnabò, G., Uva, A., Pollastrini, S., Rubagotti, C., & Bernardi, D. (2023). Supervised Clustering Loss for Clustering-Friendly Sentence Embeddings: an Application to Intent Clustering. In IJCNLP-AACL 2023 - 13th International Joint Conference on Natural Language Processing and the 3rd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics, Findings of the Association for Computational Linguistics: IJCNLP-AACL 2023 (pp. 412–430). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-ijcnlp.36

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free