Narrowing the Gap between Supervised and Unsupervised Sentence Representation Learning with Large Language Model

3Citations
Citations of this article
9Readers
Mendeley users who have this article in their library.

Abstract

Sentence Representation Learning (SRL) is a fundamental task in Natural Language Processing (NLP), with the Contrastive Learning of Sentence Embeddings (CSE) being the mainstream technique due to its superior performance. An intriguing phenomenon in CSE is the significant performance gap between supervised and unsupervised methods, with their only difference lying in the training data. Previous works attribute this performance gap to differences in two representation properties (alignment and uniformity). However, since alignment and uniformity only measure the results, they fail to answer “What aspects of the training data contribute to the performance gap?” and “How can the performance gap be narrowed?”. In this paper, we conduct empirical experiments to answer these “What” and “How” questions. We first answer the “What” question by thoroughly comparing the behavior of supervised and unsupervised CSE during their respective training processes. From the comparison, we identify the similarity pattern as a key factor to the performance gap, and introduce a metric, called Relative Fitting Difficulty (RFD), to measure the complexity of the similarity pattern. Then, based on the insights gained from the “What” question, we tackle the “How” question by increasing the pattern complexity of the training data. We achieve this by leveraging the In-Context Learning (ICL) capability of the Large Language Model (LLM) to generate data that simulates complex patterns. By utilizing the hierarchical patterns in the LLM-generated data, we effectively narrow the gap between supervised and unsupervised CSE. We release our codes and appendix at https://github.com/BDBC-KG-NLP/NGCSE.

Cite

CITATION STYLE

APA

Li, M., Zhang, R., Nie, Z., & Mao, Y. (2024). Narrowing the Gap between Supervised and Unsupervised Sentence Representation Learning with Large Language Model. In Proceedings of the AAAI Conference on Artificial Intelligence (Vol. 38, pp. 13590–13599). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v38i12.29263

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free