Lightweight and efficient neural natural language processing with quaternion networks

36Citations
Citations of this article
173Readers
Mendeley users who have this article in their library.

Abstract

Many state-of-the-art neural models for NLP are heavily parameterized and thus memory inefficient. This paper proposes a series of lightweight and memory efficient neural architectures for a potpourri of natural language processing (NLP) tasks. To this end, our models exploit computation using Quaternion algebra and hypercomplex spaces, enabling not only expressive inter-component interactions but also significantly (75%) reduced parameter size due to lesser degrees of freedom in the Hamilton product. We propose Quaternion variants of models, giving rise to new architectures such as the Quaternion attention Model and Quaternion Transformer. Extensive experiments on a battery of NLP tasks demonstrates the utility of proposed Quaternion-inspired models, enabling up to 75% reduction in parameter size without significant loss in performance.

Cite

CITATION STYLE

APA

Tay, Y., Zhang, A., Tuan, L. A., Rao, J., Zhang, S., Wang, S., … Hui, S. C. (2020). Lightweight and efficient neural natural language processing with quaternion networks. In ACL 2019 - 57th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference (pp. 1494–1503). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/p19-1145

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free