LOREM: Language-consistent Open Relation Extraction from Unstructured Text

7Citations
Citations of this article
41Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We introduce a Language-consistent multi-lingual Open Relation Extraction Model (LOREM) for finding relation tuples of any type between entities in unstructured texts. LOREM does not rely on language-specific knowledge or external NLP tools such as translators or PoS-taggers, and exploits information and structures that are consistent over different languages. This allows our model to be easily extended with only limited training efforts to new languages, but also provides a boost to performance for a given single language. An extensive evaluation performed on 5 languages shows that LOREM outperforms state-of-the-art mono-lingual and cross-lingual open relation extractors. Moreover, experiments on languages with no or only little training data indicate that LOREM generalizes to other languages than the languages that it is trained on.

Cite

CITATION STYLE

APA

Harting, T., Mesbah, S., & Lofi, C. (2020). LOREM: Language-consistent Open Relation Extraction from Unstructured Text. In The Web Conference 2020 - Proceedings of the World Wide Web Conference, WWW 2020 (pp. 1830–1838). Association for Computing Machinery, Inc. https://doi.org/10.1145/3366423.3380252

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free