Abstract
Modern virtual assistants are trained to classify customer requests into a taxonomy of predesigned intents. Requests that fall outside of this taxonomy, however, are often unhandled and need to be clustered to define new experiences. Recently, state-of-the-art results in intent clustering were achieved by training a neural network with a latent structured prediction loss. Unfortunately, though, this new approach suffers from a quadratic bottleneck as it requires to compute a joint embedding representation for all pairs of utterances to cluster. To overcome this limitation, we instead cast the problem into a representation learning task, and we adapt the latent structured prediction loss to fine-tune sentence encoders, thus making it possible to obtain clustering-friendly single-sentence embeddings. Our experiments show that the supervised clustering loss returns state-of-the-art results in terms of clustering accuracy and adjusted mutual information.
Cite
CITATION STYLE
Barnabò, G., Uva, A., Pollastrini, S., Rubagotti, C., & Bernardi, D. (2023). Supervised Clustering Loss for Clustering-Friendly Sentence Embeddings: an Application to Intent Clustering. In IJCNLP-AACL 2023 - 13th International Joint Conference on Natural Language Processing and the 3rd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics, Findings of the Association for Computational Linguistics: IJCNLP-AACL 2023 (pp. 412–430). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-ijcnlp.36
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.