Abstract
Low-shot relation extraction (RE) aims to recognize novel relations with very few or even no samples, which is critical in real scenario application. Few-shot and zero-shot RE are two representative low-shot RE tasks, which seem to be with similar target but require totally different underlying abilities. In this paper, we propose Multi-Choice Matching Networks to unify low-shot relation extraction. To fill in the gap between zero-shot and few-shot RE, we propose the triplet-paraphrase meta-training, which leverages triplet paraphrase to pre-train zero-shot label matching ability and uses meta-learning paradigm to learn few-shot instance summarizing ability. Experimental results on three different low-shot RE tasks show that the proposed method outperforms strong baselines by a large margin, and achieve the best performance on few-shot RE leaderboard.
Cite
CITATION STYLE
Liu, F., Lin, H., Han, X., Cao, B., & Sun, L. (2022). Pre-training to Match for Unified Low-shot Relation Extraction. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 1, pp. 5785–5795). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.acl-long.397
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.