Question answering on freebase via relation extraction and textual evidence

231Citations
Citations of this article
371Readers
Mendeley users who have this article in their library.

Abstract

Existing knowledge-based question answering systems often rely on small annotated training data. While shallow methods like relation extraction are robust to data scarcity, they are less expressive than the deep meaning representation methods like semantic parsing, thereby failing at answering questions involving multiple constraints. Here we alleviate this problem by empowering a relation extraction method with additional evidence from Wikipedia. We first present a neural network based relation extractor to retrieve the candidate answers from Freebase, and then infer over Wikipedia to validate these answers. Experiments on the WebQuestions question answering dataset show that our method achieves an F1 of 53.3%, a substantial improvement over the state-of-the-art.

Cite

CITATION STYLE

APA

Xu, K., Reddy, S., Feng, Y., Huang, S., & Zhao, D. (2016). Question answering on freebase via relation extraction and textual evidence. In 54th Annual Meeting of the Association for Computational Linguistics, ACL 2016 - Long Papers (Vol. 4, pp. 2326–2336). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/p16-1220

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free