Generating Labeled Data for Relation Extraction: A Meta Learning Approach with Joint GPT-2 Training

1Citations
Citations of this article
16Readers
Mendeley users who have this article in their library.

Abstract

Relation Extraction (RE) is the task of identifying semantic relation between real-world entities mentioned in text. Despite significant progress in RE research, a remaining challenge for RE concerns the lack of training data for data-hungry deep learning models. Cost of annotation and difficulty of the task are among hindrance to collect a large-scale RE dataset in different domains. To address this limitation, we propose a novel framework to automatically generate labeled data for RE. Our framework presents the pre-trained language model GPT-2 for data generation. In addition, to optimize the generated samples for an RE model, we introduce a meta learning approach to allow the GPT-2 model to be updated during the training process for RE. In particular, to leverage the feedback from the RE model to improve the data generation from GPT-2, we propose a novel reward function to update the GPT-2 model with REINFORCE, seeking to promote the similarity of the RE loss function's gradients computed for generated data and a meta development set. We conduct extensive experiments on two benchmark datasets to produce state-of-the-art performance for RE.

Cite

CITATION STYLE

APA

Veyseh, A. P. B., Dernoncourt, F., Min, B., & Nguyen, T. H. (2023). Generating Labeled Data for Relation Extraction: A Meta Learning Approach with Joint GPT-2 Training. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 11466–11478). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-acl.727

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free