KEML: A Knowledge-Enriched Meta-Learning Framework for Lexical Relation Classification

12Citations
Citations of this article
35Readers
Mendeley users who have this article in their library.

Abstract

Lexical relations describe how concepts are semantically related, in the form of relation triples. The accurate prediction of lexical relations between concepts is challenging, due to the sparsity of patterns indicating the existence of such relations. We propose the Knowledge-Enriched Meta-Learning (KEML) framework to address lexical relation classification. In KEML, the LKB-BERT (Lexical Knowledge Base-BERT) model is first presented to learn concept representations from text corpora, with rich lexical knowledge injected by distant supervision. A probabilistic distribution of auxiliary tasks is defined to increase the model's ability to recognize different types of lexical relations. We further propose a neural classifier integrated with special relation recognition cells, in order to combine meta-learning over the auxiliary task distribution and supervised learning for LRC. Experiments over multiple datasets show KEML outperforms state-of-the-art methods.

Cite

CITATION STYLE

APA

Wang, C., Qiu, M., Huang, J., & He, X. (2021). KEML: A Knowledge-Enriched Meta-Learning Framework for Lexical Relation Classification. In 35th AAAI Conference on Artificial Intelligence, AAAI 2021 (Vol. 15, pp. 13924–13932). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v35i15.17640

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free