Large-scale learning of relation-extraction rules with distant supervision from the web

55Citations
Citations of this article
83Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

We present a large-scale relation extraction (RE) system which learns grammar-based RE rules from the Web by utilizing large numbers of relation instances as seed. Our goal is to obtain rule sets large enough to cover the actual range of linguistic variation, thus tackling the long-tail problem of real-world applications. A variant of distant supervision learns several relations in parallel, enabling a new method of rule filtering. The system detects both binary and n-ary relations. We target 39 relations from Freebase, for which 3M sentences extracted from 20M web pages serve as the basis for learning an average of 40K distinctive rules per relation. Employing an efficient dependency parser, the average run time for each relation is only 19 hours. We compare these rules with ones learned from local corpora of different sizes and demonstrate that the Web is indeed needed for a good coverage of linguistic variation. © 2012 Springer-Verlag Berlin Heidelberg.

Cite

CITATION STYLE

APA

Krause, S., Li, H., Uszkoreit, H., & Xu, F. (2012). Large-scale learning of relation-extraction rules with distant supervision from the web. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 7649 LNCS, pp. 263–278). Springer Verlag. https://doi.org/10.1007/978-3-642-35176-1_17

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free