A typical product or place often has hundreds of reviews, and summarization of these texts is a challenging problem. Recent progress on abstractive summarization in domains such as news has been driven by supervised systems trained on hundreds of thousands of news articles paired with human-written summaries. However for opinion texts, such large scale datasets are rarely available. Unsupervised methods, self-training, and few-shot learning approaches bridge that gap. In this work, we present a novel self-training approach, OPINESUM for abstractive opinion summarization. The self-training summaries in this approach are built automatically using a novel application of textual entailment and capture the consensus of opinions across the various reviews for an item. This method can be used to obtain silver-standard summaries on a large scale and train both unsupervised and few-shot abstractive summarization systems. OPINESUM outperforms strong peer systems in both settings.
CITATION STYLE
Louis, A., & Maynez, J. (2023). OPINESUM: Entailment-based self-training for abstractive opinion summarization. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 10774–10790). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-acl.686
Mendeley helps you to discover research relevant for your work.