Sentence meta-embeddings for unsupervised semantic textual similarity

16Citations
Citations of this article
116Readers
Mendeley users who have this article in their library.

Abstract

We address the task of unsupervised Semantic Textual Similarity (STS) by ensembling diverse pre-trained sentence encoders into sentence meta-embeddings. We apply, extend and evaluate different meta-embedding methods from the word embedding literature at the sentence level, including dimensionality reduction (Yin and Schütze, 2016), generalized Canonical Correlation Analysis (Rastogi et al., 2015) and cross-view auto-encoders (Bollegala and Bao, 2018). Our sentence meta-embeddings set a new unsupervised State of The Art (SoTA) on the STS Benchmark and on the STS12-STS16 datasets, with gains of between 3.7% and 6.4% Pearson's r over single-source systems.

Cite

CITATION STYLE

APA

Poerner, N., Waltinger, U., & Schütze, H. (2020). Sentence meta-embeddings for unsupervised semantic textual similarity. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 7027–7034). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.acl-main.628

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free