Tensorized embedding layers

27Citations
Citations of this article
78Readers
Mendeley users who have this article in their library.

Abstract

The embedding layers transforming input words into real vectors are the key components of deep neural networks used in natural language processing. However, when the vocabulary is large, the corresponding weight matrices can be enormous, which precludes their deployment in a limited resource setting. We introduce a novel way of parameterizing embedding layers based on the Tensor Train decomposition, which allows compressing the model significantly at the cost of a negligible drop or even a slight gain in performance. We evaluate our method on a wide range of benchmarks in natural language processing and analyze the trade-off between performance and compression ratios for a wide range of architectures, from MLPs to LSTMs and Transformers.

Cite

CITATION STYLE

APA

Hrinchuk, O., Khrulkov, V., Mirvakhabova, L., Orlova, E., & Oseledets, I. (2020). Tensorized embedding layers. In Findings of the Association for Computational Linguistics Findings of ACL: EMNLP 2020 (pp. 4847–4860). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.findings-emnlp.436

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free