Uniform Complexity for Text Generation

0Citations
Citations of this article
11Readers
Mendeley users who have this article in their library.

Abstract

Large language models (LLMs) have shown promising results in a wide array of generative NLP tasks, such as summarization and machine translation. In the context of narrative generation, however, existing models still do not capture factors that contribute to producing consistent text. For instance, it is logical that a piece of text or a story should be uniformly readable throughout and that this form of complexity should be controllable. As such, if the complexity of an input text prompt is rated first-grade reading level in the Flesch Reading Ease test, then the generated text continuing the plot should also be within this range of complexity. With this in mind, we introduce Uniform Complexity for Text Generation (UCTG), a new benchmark test which raises the challenge of making generative models observe uniform linguistic properties with respect to prompts. We experiment with over 150+ linguistically and cognitively motivated features for evaluating text complexity in humans and generative models. From our results, we find that models such as GPT-2 struggle to preserve the complexity of input prompts used in its generations, even if finetuned with professionally written texts.

Cite

CITATION STYLE

APA

Imperial, J. M., & Madabushi, H. T. (2023). Uniform Complexity for Text Generation. In Findings of the Association for Computational Linguistics: EMNLP 2023 (pp. 12025–12046). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-emnlp.805

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free