What Has Been Enhanced in my Knowledge-Enhanced Language Model?

0Citations
Citations of this article
22Readers
Mendeley users who have this article in their library.
Get full text

Abstract

A number of knowledge integration (KI) methods have recently been proposed to incorporate external knowledge into pretrained language models (LMs). Even though knowledge-enhanced LMs outperform base LMs on knowledge-intensive tasks, the inner-workings of these KI methods are not well-understood. For instance, it is unclear which knowledge is effectively integrated into knowledge-enhanced LMs and which is not; and if such integration leads to catastrophic forgetting of already learned knowledge. We show that existing model interpretation methods such as linear probes and prompts have some key limitations in answering these questions. We revisit KI from an information-theoretic view and propose a new theoretically sound probe called Graph Convolution Simulator (GCS) for KI interpretation. GCS uses graph attention on the corresponding knowledge graph for interpretation. In our experiments we verify that GCS can provide reasonable interpretation results for two well-known knowledge-enhanced LMs: ERNIE and K-Adapter. We also find that only a marginal amount of knowledge is successfully integrated in these models, and simply increasing the size of the KI corpus may not lead to better knowledge-enhanced LMs.

Cite

CITATION STYLE

APA

Hou, Y., Fu, G., & Sachan, M. (2022). What Has Been Enhanced in my Knowledge-Enhanced Language Model? In Findings of the Association for Computational Linguistics: EMNLP 2022 (pp. 1417–1438). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.findings-emnlp.475

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free