A Case Study on Learning a Unified Encoder of Relations

3Citations
Citations of this article
69Readers
Mendeley users who have this article in their library.

Abstract

Typical relation extraction models are trained on a single corpus annotated with a pre-defined relation schema. An individual corpus is often small, and the models may often be biased or overfitted to the corpus. We hypothesize that we can learn a better representation by combining multiple relation datasets. We attempt to use a shared encoder to learn the unified feature representation and to augment it with regularization by adversarial training. The additional corpora feeding the encoder can help to learn a better feature representation layer even though the relation schemas are different. We use ACE05 and ERE datasets as our case study for experiments. The multi-task model obtains significant improvement on both datasets.

Cite

CITATION STYLE

APA

Fu, L., Min, B., Nguyen, T. H., & Grishman, R. (2018). A Case Study on Learning a Unified Encoder of Relations. In 4th Workshop on Noisy User-Generated Text, W-NUT 2018 - Proceedings of the Workshop (pp. 202–207). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/w18-6126

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free