Unsupervised Sentence Representation via Contrastive Learning with Mixing Negatives

42Citations
Citations of this article
25Readers
Mendeley users who have this article in their library.

Abstract

Unsupervised sentence representation learning is a fundamental problem in natural language processing. Recently, contrastive learning has made great success on this task. Existing constrastive learning based models usually apply random sampling to select negative examples for training. Previous work in computer vision has shown that hard negative examples help contrastive learning to achieve faster convergency and better optimization for representation learning. However, the importance of hard negatives in contrastive learning for sentence representation is yet to be explored. In this study, we prove that hard negatives are essential for maintaining strong gradient signals in the training process while random sampling negative examples is ineffective for sentence representation. Accordingly, we present a contrastive model, MixCSE, that extends the current state-of-the-art Sim- CSE by continually constructing hard negatives via mixing both positive and negative features. The superior performance of the proposed approach is demonstrated via empirical studies on Semantic Textual Similarity datasets and Transfer task datasets.

Cite

CITATION STYLE

APA

Zhang, Y., Zhang, R., Mensah, S., Liu, X., & Mao, Y. (2022). Unsupervised Sentence Representation via Contrastive Learning with Mixing Negatives. In Proceedings of the 36th AAAI Conference on Artificial Intelligence, AAAI 2022 (Vol. 36, pp. 11730–11738). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v36i10.21428

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free