Relying on Discourse Analysis to Answer Complex Questions by Neural Machine Reading Comprehension

2Citations
Citations of this article
42Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Machine reading comprehension (MRC) is one of the most challenging tasks in natural language processing domain. Recent state-of-the-art results for MRC have been achieved with the pre-trained language models, such as BERT and its modifications. Despite the high performance of these models, they still suffer from the inability to retrieve correct answers from the detailed and lengthy passages. In this work, we introduce a novel scheme for incorporating the discourse structure of the text into a self-attention network, and, thus, enrich the embedding obtained from the standard BERT encoder with the additional linguistic knowledge. We also investigate the influence of different types of linguistic information on the model's ability to answer complex questions that require deep understanding of the whole text. Experiments performed on the SQuAD benchmark and more complex question answering datasets have shown that linguistic enhancing boosts the performance of the standard BERT model significantly.

Cite

CITATION STYLE

APA

Galitsky, B., Ilvovsky, D., & Goncharova, E. (2021). Relying on Discourse Analysis to Answer Complex Questions by Neural Machine Reading Comprehension. In International Conference Recent Advances in Natural Language Processing, RANLP (pp. 444–453). Incoma Ltd. https://doi.org/10.26615/978-954-452-072-4_051

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free