Gradient Imitation Reinforcement Learning for Low Resource Relation Extraction

52Citations
Citations of this article
89Readers
Mendeley users who have this article in their library.

Abstract

Low-resource Relation Extraction (LRE) aims to extract relation facts from limited labeled corpora when human annotation is scarce. Existing works either utilize self-training scheme to generate pseudo labels that will cause the gradual drift problem, or leverage meta-learning scheme which does not solicit feedback explicitly. To alleviate selection bias due to the lack of feedback loops in existing LRE learning paradigms, we developed a Gradient Imitation Reinforcement Learning method to encourage pseudo label data to imitate the gradient descent direction on labeled data and bootstrap its optimization capability through trial and error. We also propose a framework called GradLRE, which handles two major scenarios in low-resource relation extraction. Besides the scenario where unlabeled data is sufficient, GradLRE handles the situation where no unlabeled data is available, by exploiting a contextualized augmentation method to generate data. Experimental results on two public datasets demonstrate the effectiveness of GradLRE on low resource relation extraction when comparing with baselines. Source code is available.

Cite

CITATION STYLE

APA

Hu, X., Zhang, C., Yang, Y., Li, X., Lin, L., Wen, L., & Yu, P. S. (2021). Gradient Imitation Reinforcement Learning for Low Resource Relation Extraction. In EMNLP 2021 - 2021 Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 2737–2746). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.emnlp-main.216

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free