Abstract
Open-domain textual question answering (QA), which aims to answer questions from large data sources like Wikipedia or the web, has gained wide attention in recent years. Recent advancements in open-domain textual QA are mainly due to the significant developments of deep learning techniques, especially machine reading comprehension and neural-network-based information retrieval, which allows the models to continuously refresh state-of-the-art performances. However, a comprehensive review of existing approaches and recent trends is lacked in this field. To address this issue, we present a thorough survey to explicitly give the task scope of open-domain textual QA, overview recent key advancements on deep learning based open-domain textual QA, illustrate the models and acceleration methods in detail, and introduce open-domain textual QA datasets and evaluation metrics. Finally, we summary the models, discuss the limitations of existing works and potential future research directions.
Author supplied keywords
Cite
CITATION STYLE
Huang, Z., Xu, S., Hu, M., Wang, X., Qiu, J., Fu, Y., … Wang, C. (2020). Recent Trends in Deep Learning Based Open-Domain Textual Question Answering Systems. IEEE Access, 8, 94341–94356. https://doi.org/10.1109/ACCESS.2020.2988903
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.