Can Machines Learn to Comprehend Scientific Literature?

8Citations
Citations of this article
36Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

To measure the ability of a machine to understand professional-level scientific articles, we construct a scientific question answering task called PaperQA. The PaperQA task is based on more than 80 000 'fill-in-the-blank' type questions on articles from reputed scientific journals such as Nature and Science. We perform fine-grained linguistic analysis and evaluation to compare PaperQA and other conventional question and answering (QA) tasks on general literature (e.g., books, news articles, and Wikipedia texts). The results indicate that the PaperQA task is the most difficult QA task for both humans (lay people) and machines (deep-learning models). Moreover, humans generally outperform machines in conventional QA tasks, but we found that advanced deep-learning models outperform humans by 3%-13% on average in the PaperQA task. The PaperQA dataset used in this paper is publicly available at http://dmis.korea.ac.kr/downloads?id=PaperQA.

Cite

CITATION STYLE

APA

Park, D., Choi, Y., Kim, D., Yu, M., Kim, S., & Kang, J. (2019). Can Machines Learn to Comprehend Scientific Literature? IEEE Access, 7, 16246–16256. https://doi.org/10.1109/ACCESS.2019.2891666

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free