Abstract
Style transfer is the task of paraphrasing text into a target-style domain while retaining the content. Unsupervised approaches mainly focus on training a generator to rewrite input sentences. In this work, we assume that text styles are determined by only a small proportion of words; therefore, rewriting sentences via generative models may be unnecessary. As an alternative, we consider style transfer as a sequence tagging task. Specifically, we use edit operations (i.e., deletion, insertion and substitution) to tag words in an input sentence. We train a classifier and a language model to score tagged sequences and build a conditional random field. Finally, the optimal path in the conditional random field is used as the output. The results of experiments comparing models indicate that our proposed model exceeds end-to-end baselines in terms of accuracy on both sentiment and style transfer tasks with comparable or better content preservation.
Cite
CITATION STYLE
Yang, S. (2022). Tagging Without Rewriting: A Probabilistic Model for Unpaired Sentiment and Style Transfer. In WASSA 2022 - 12th Workshop on Computational Approaches to Subjectivity, Sentiment and Social Media Analysis, Proceedings of the Workshop (pp. 293–303). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.wassa-1.33
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.