Abstract
Automated storytelling has long captured the attention of researchers for the ubiquity of narratives in everyday life. However, it is challenging to maintain coherence and stay on-topic toward a specific ending when generating narratives with neural language models. In this paper, we introduce Story generation with Reader Models (StoRM), a framework in which a reader model is used to reason about the story should progress. A reader model infers what a human reader believes about the concepts, entities, and relations about the fictional story world. We show how an explicit reader model represented as a knowledge graph affords story coherence and provides controllability in the form of achieving a given story world state goal. Experiments show that our model produces significantly more coherent and on-topic stories, outperforming baselines in dimensions including plot plausibility and staying on topic.
Cite
CITATION STYLE
Peng, X., Xie, K., Alabdulkarim, A., Kayam, H., Dani, S., & Riedl, M. O. (2022). Guiding Neural Story Generation with Reader Models. In Findings of the Association for Computational Linguistics: EMNLP 2022 (pp. 7116–7140). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.findings-emnlp.391
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.