Abstract
We study the problem of domain adaptation for neural abstractive summarization. We make initial efforts in investigating what information can be transferred to a new domain. Experimental results on news stories and opinion articles indicate that neural summarization model benefits from pre-training based on extractive summaries. We also find that the combination of in-domain and out-of-domain setup yields better summaries when in-domain data is insufficient. Further analysis shows that, the model is capable to select salient content even trained on out-of-domain data, but requires in-domain data to capture the style for a target domain.
Cite
CITATION STYLE
Hua, X., & Wang, L. (2017). A pilot study of domain adaptation effect for neural abstractive summarization. In EMNLP 2017 - Workshop on New Frontiers in Summarization, NFiS 2017 - Workshop Proceedings (pp. 100–106). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/w17-4513
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.