Relation extraction is an important task in structuring content of text data, and becomes especially challenging when learning with weak supervision-where only a limited number of labeled sentences are given and a large number of unlabeled sentences are available. Most existing work exploits unlabeled data based on the ideas of self-training (i.e., bootstrapping a model) and self-ensembling (e.g., ensembling multiple model variants). However, these methods either suffer from the issue of semantic drift, or do not fully capture the problem characteristics of relation extraction. In this paper, we leverage a key insight that retrieving sentences expressing a relation is a dual task of predicting the relation label for a given sentence-two tasks are complementary to each other and can be optimized jointly for mutual enhancement. To model this intuition, we propose DualRE, a principled framework that introduces a retrieval module which is jointly trained with the original relation prediction module. In this way, high-quality samples selected by the retrieval module from unlabeled data can be used to improve the prediction module, and vice versa. Experimental results1 on two public datasets as well as case studies demonstrate the effectiveness of the DualRE approach.
CITATION STYLE
Lin, H., Qu, M., Yan, J., & Ren, X. (2019). Learning dual retrieval module for semi-supervised relation extraction. In The Web Conference 2019 - Proceedings of the World Wide Web Conference, WWW 2019 (pp. 1073–1083). Association for Computing Machinery, Inc. https://doi.org/10.1145/3308558.3313573
Mendeley helps you to discover research relevant for your work.