RobustEmbed: Robust Sentence Embeddings Using Self-Supervised Contrastive Pre-Training

1Citations
Citations of this article
10Readers
Mendeley users who have this article in their library.

Abstract

Pre-trained language models (PLMs) have demonstrated exceptional performance across a wide range of natural language processing tasks. The utilization of PLM-based sentence embeddings enables the generation of contextual representations that capture rich semantic information. However, despite their success with unseen samples, current PLM-based representations suffer from poor robustness in adversarial settings. In this paper, we propose RobustEmbed, a self-supervised sentence embedding framework that enhances both generalization and robustness in various text representation tasks and against a diverse set of adversarial attacks. By generating high-risk adversarial perturbations to promote higher invariance in the embedding space and leveraging the perturbation within a novel contrastive objective approach, RobustEmbed effectively learns high-quality sentence embeddings. Our extensive experiments validate the superiority of RobustEmbed over the state-of-the-art self-supervised representations in adversarial settings, while also showcasing relative improvements in seven semantic textual similarity (STS) tasks and six transfer tasks. Specifically, our framework achieves a significant reduction in attack success rate from 75.51% to 39.62% for the BERTAttack attack technique, along with enhancements of 1.20% and 0.40% in STS tasks and transfer tasks, respectively.

Cite

CITATION STYLE

APA

Asl, J. R., Blanco, E., & Takabi, D. (2023). RobustEmbed: Robust Sentence Embeddings Using Self-Supervised Contrastive Pre-Training. In Findings of the Association for Computational Linguistics: EMNLP 2023 (pp. 4587–4603). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-emnlp.305

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free