CoLAKE: Contextualized Language and Knowledge Embedding

139Citations
Citations of this article
249Readers
Mendeley users who have this article in their library.

Abstract

With the emerging branch of incorporating factual knowledge into pre-trained language models such as BERT, most existing models consider shallow, static, and separately pre-trained entity embeddings, which limits the performance gains of these models. Few works explore the potential of deep contextualized knowledge representation when injecting knowledge. In this paper, we propose the Contextualized Language and Knowledge Embedding (CoLAKE), which jointly learns contextualized representation for both language and knowledge with the extended MLM objective. Instead of injecting only entity embeddings, CoLAKE extracts the knowledge context of an entity from large-scale knowledge bases. To handle the heterogeneity of knowledge context and language context, we integrate them in a unified data structure, word-knowledge graph (WK graph). CoLAKE is pre-trained on large-scale WK graphs with the modified Transformer encoder. We conduct experiments on knowledge-driven tasks, knowledge probing tasks, and language understanding tasks. Experimental results show that CoLAKE outperforms previous counterparts on most of the tasks. Besides, CoLAKE achieves surprisingly high performance on our synthetic task called word-knowledge graph completion, which shows the superiority of simultaneously contextualizing language and knowledge representation.

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Sun, T., Shao, Y., Qiu, X., Guo, Q., Hu, Y., Huang, X., & Zhang, Z. (2020). CoLAKE: Contextualized Language and Knowledge Embedding. In COLING 2020 - 28th International Conference on Computational Linguistics, Proceedings of the Conference (pp. 3660–3670). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.coling-main.327

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 80

75%

Researcher 19

18%

Lecturer / Post doc 5

5%

Professor / Associate Prof. 3

3%

Readers' Discipline

Tooltip

Computer Science 100

88%

Linguistics 8

7%

Engineering 3

3%

Business, Management and Accounting 3

3%

Save time finding and organizing research with Mendeley

Sign up for free