CopyBERT: A unified approach to question generation with self-attention

10Citations
Citations of this article
97Readers
Mendeley users who have this article in their library.

Abstract

Contextualized word embeddings provide better initialization for neural networks that deal with various natural language understanding (NLU) tasks including question answering (QA) and more recently, question generation (QG). Apart from providing meaningful word representations, pre-trained transformer models, such as BERT also provide self-attentions which encode syntactic information that can be probed for dependency parsing and POS-tagging. In this paper, we show that the information from self-attentions of BERT are useful for language modeling of questions conditioned on paragraph and answer phrases. To control the attention span, we use semi-diagonal mask and utilize a shared model for encoding and decoding, unlike sequence-to-sequence. We further employ copy mechanism over self-attentions to achieve state-of-the-art results for question generation on SQuAD dataset.

References Powered by Scopus

SQuad: 100,000+ questions for machine comprehension of text

4027Citations
2234Readers

Incorporating copying mechanism in sequence-to-sequence learning

995Citations
1245Readers

CoQA: A Conversational Question Answering Challenge

747Citations
686Readers

Cited by Powered by Scopus

Question Generation for Reading Comprehension Assessment by Modeling How and What to Ask

16Citations
73Readers

AutoEQA: Auto-Encoding Questions for Extractive Question Answering

2Citations
50Readers

Question Generation via Generative Adversarial Networks

1Citations
2Readers
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Varanasi, S., Amin, S., & Neumann, G. (2020). CopyBERT: A unified approach to question generation with self-attention. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 25–31). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.nlp4convai-1.3

Readers over time

‘20‘21‘22‘23‘24‘2507142128

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 36

72%

Researcher 10

20%

Lecturer / Post doc 4

8%

Readers' Discipline

Tooltip

Computer Science 45

80%

Linguistics 7

13%

Social Sciences 2

4%

Engineering 2

4%

Save time finding and organizing research with Mendeley

Sign up for free
0