A Global Past-Future Early Exit Method for Accelerating Inference of Pre-trained Language Models

39Citations
Citations of this article
63Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Early exit mechanism aims to accelerate the inference speed of large-scale pre-trained language models. The essential idea is to exit early without passing through all the inference layers at the inference stage. To make accurate predictions for downstream tasks, the hierarchical linguistic information embedded in all layers should be jointly considered. However, much of the research up to now has been limited to use local representations of the exit layer. Such treatment inevitably loses information of the unused past layers as well as the high-level features embedded in future layers, leading to sub-optimal performance. To address this issue, we propose a novel Past-Future method to make comprehensive predictions from a global perspective. We first take into consideration all the linguistic information embedded in the past layers and further engage the future information which is originally inaccessible for predictions. Extensive experiments demonstrate that our method outperforms previous early exit methods by a large margin, yielding better and robust performance.

Cite

CITATION STYLE

APA

Liao, K., Zhang, Y., Ren, X., Su, Q., Sun, X., & He, B. (2021). A Global Past-Future Early Exit Method for Accelerating Inference of Pre-trained Language Models. In NAACL-HLT 2021 - 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Conference (pp. 2013–2023). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.naacl-main.162

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free