Enhancing Chinese Pre-trained Language Model via Heterogeneous Linguistics Graph

7Citations
Citations of this article
42Readers
Mendeley users who have this article in their library.

Abstract

Chinese pre-trained language models usually exploit contextual character information to learn representations, while ignoring the linguistics knowledge, e.g., word and sentence information. Hence, we propose a task-free enhancement module termed as Heterogeneous Linguistics Graph (HLG) to enhance Chinese pre-trained language models by integrating linguistics knowledge. Specifically, we construct a hierarchical heterogeneous graph to model the characteristics linguistics structure of Chinese language, and conduct a graph-based method to summarize and concretize information on different granularities of Chinese linguistics hierarchies. Experimental results demonstrate our model has the ability to improve the performance of vanilla BERT, BERTwwm and ERNIE 1.0 on 6 natural language processing tasks with 10 benchmark datasets. Further, the detailed experimental analyses have proven that this kind of modelization achieves more improvements compared with previous strong baseline MWA. Meanwhile, our model introduces far fewer parameters (about half of MWA) and the training/inference speed is about 7x faster than MWA. Our code and processed datasets are available at https://github.com/lsvih/HLG.

Cite

CITATION STYLE

APA

Li, Y., Cao, J., Cong, X., Zhang, Z., Yu, B., Zhu, H., & Liu, T. (2022). Enhancing Chinese Pre-trained Language Model via Heterogeneous Linguistics Graph. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 1, pp. 1986–1996). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.acl-long.140

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free