Semantics of the unwritten: The effect of end of paragraph and sequence tokens on text generation with GPT2

4Citations
Citations of this article
47Readers
Mendeley users who have this article in their library.

Abstract

The semantics of a text is manifested not only by what is read, but also by what is not read. In this article, we will study how the implicit “not read” information such as end-of-paragraph (EOP) and end-of-sequence (EOS) affect the quality of text generation. Specifically, we find that the pre-trained language model GPT2 can generate better continuations by learning to generate the EOP in the fine-tuning stage. Experimental results on English story generation show that EOP can lead to higher BLEU score and lower EOS perplexity. We also conduct experiments on a self-collected Chinese essay dataset with Chinese-GPT2, a character level LM without EOP or EOS during pre-training. Experimental results show that the Chinese GPT2 can generate better essay endings with EOP. Our code is available on GitHub.

Cite

CITATION STYLE

APA

Bai, H., Shi, P., Lin, J., Tan, L., Xiong, K., Gao, W., … Li, M. (2021). Semantics of the unwritten: The effect of end of paragraph and sequence tokens on text generation with GPT2. In ACL-IJCNLP 2021 - 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, Proceedings of the Student Research Workshop (pp. 148–162). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.acl-srw.16

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free