Zero-Shot Aspect-Based Scientific Document Summarization using Self-Supervised Pre-training

5Citations
Citations of this article
37Readers
Mendeley users who have this article in their library.

Abstract

We study the zero-shot setting for the aspect-based scientific document summarization task. Summarizing scientific documents with respect to an aspect can remarkably improve document assistance systems and readers experience. However, existing large-scale datasets contain a limited variety of aspects, causing summarization models to over-fit to a small set of aspects and a specific domain. We establish baseline results in zero-shot performance (over unseen aspects and the presence of domain shift), paraphrasing, leave-one-out, and limited supervised samples experimental setups. We propose a self-supervised pre-training approach to enhance the zero-shot performance. We leverage the PubMed structured abstracts to create a biomedical aspect-based summarization dataset. Experimental results on the PubMed and FacetSum aspect-based datasets show promising performance when the model is pre-trained using unlabelled in-domain data.

Cite

CITATION STYLE

APA

Soleimani, A., Nikoulina, V., Favre, B., & Ait-Mokhtar, S. (2022). Zero-Shot Aspect-Based Scientific Document Summarization using Self-Supervised Pre-training. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 49–62). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.bionlp-1.5

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free