Sentiment-Aware Word and Sentence Level Pre-training for Sentiment Analysis

13Citations
Citations of this article
27Readers
Mendeley users who have this article in their library.

Abstract

Most existing pre-trained language representation models (PLMs) are sub-optimal in sentiment analysis tasks, as they capture the sentiment information from word-level while under-considering sentence-level information. In this paper, we propose SentiWSP, a novel Sentiment-aware pre-trained language model with combined Word-level and Sentence-level Pre-training tasks. The word level pre-training task detects replaced sentiment words, via a generator-discriminator framework, to enhance the PLM's knowledge about sentiment words. The sentence level pre-training task further strengthens the discriminator via a contrastive learning framework, with similar sentences as negative samples, to encode sentiments in a sentence. Extensive experimental results show that SentiWSP achieves new state-of-the-art performance on various sentence-level and aspect-level sentiment classification benchmarks. We have made our code and model publicly available at https://github.com/XMUDM/SentiWSP.

Cite

CITATION STYLE

APA

Fan, S., Lin, C., Li, H., Lin, Z., Su, J., Zhang, H., … Duan, N. (2022). Sentiment-Aware Word and Sentence Level Pre-training for Sentiment Analysis. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022 (pp. 4984–4994). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.emnlp-main.332

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free