Evaluation metrics for machine reading comprehension: Prerequisite skills and readability

36Citations
Citations of this article
171Readers
Mendeley users who have this article in their library.

Abstract

Knowing the quality of reading comprehension (RC) datasets is important for the development of natural-language understanding systems. In this study, two classes of metrics were adopted for evaluating RC datasets: prerequisite skills and readability. We applied these classes to six existing datasets, including MCTest and SQuAD, and highlighted the characteristics of the datasets according to each metric and the correlation between the two classes. Our dataset analysis suggests that the readability of RC datasets does not directly affect the question difficulty and that it is possible to create an RC dataset that is easy to read but difficult to answer.

Cite

CITATION STYLE

APA

Sugawara, S., Kido, Y., Yokono, H., & Aizawa, A. (2017). Evaluation metrics for machine reading comprehension: Prerequisite skills and readability. In ACL 2017 - 55th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference (Long Papers) (Vol. 1, pp. 806–817). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/P17-1075

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free