Typical relation extraction models are trained on a single corpus annotated with a pre-defined relation schema. An individual corpus is often small, and the models may often be biased or overfitted to the corpus. We hypothesize that we can learn a better representation by combining multiple relation datasets. We attempt to use a shared encoder to learn the unified feature representation and to augment it with regularization by adversarial training. The additional corpora feeding the encoder can help to learn a better feature representation layer even though the relation schemas are different. We use ACE05 and ERE datasets as our case study for experiments. The multi-task model obtains significant improvement on both datasets.
CITATION STYLE
Fu, L., Min, B., Nguyen, T. H., & Grishman, R. (2018). A Case Study on Learning a Unified Encoder of Relations. In 4th Workshop on Noisy User-Generated Text, W-NUT 2018 - Proceedings of the Workshop (pp. 202–207). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/w18-6126
Mendeley helps you to discover research relevant for your work.