EdgeFormer: A Parameter-Efficient Transformer for On-Device Seq2seq Generation

9Citations
Citations of this article
43Readers
Mendeley users who have this article in their library.

Abstract

We introduce EDGEFORMER - a parameter-efficient Transformer for on-device seq2seq generation under the strict computation and memory constraints. Compared with the previous parameter-efficient Transformers, EDGEFORMER applies two novel principles for cost-effective parameterization, allowing it to perform better given the same parameter budget; moreover, EDGEFORMER is further enhanced by layer adaptation innovation that is proposed for improving the network with shared layers. Extensive experiments show EDGEFORMER can effectively outperform previous parameter-efficient Transformer baselines and achieve competitive results under both the computation and memory constraints. Given the promising results, we release EDGELM - the pretrained version of EDGEFORMER, which is the first publicly available pretrained on-device seq2seq model that can be easily fine-tuned for seq2seq tasks with strong results, facilitating on-device seq2seq generation in practice.

References Powered by Scopus

Convolutional sequence to sequence learning

2152Citations
N/AReaders
Get full text

The CoNLL-2014 shared task on grammatical error correction

438Citations
N/AReaders
Get full text

Learning to ask: Neural question generation for reading comprehension

435Citations
N/AReaders
Get full text

Cited by Powered by Scopus

Efficient Methods for Natural Language Processing: A Survey

46Citations
N/AReaders
Get full text

Speculative Decoding: Exploiting Speculative Execution for Accelerating Seq2seq Generation

17Citations
N/AReaders
Get full text

Exploring All-In-One Knowledge Distillation Framework for Neural Machine Translation

6Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Ge, T., Chen, S. Q., & Wei, F. (2022). EdgeFormer: A Parameter-Efficient Transformer for On-Device Seq2seq Generation. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022 (pp. 10786–10798). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.emnlp-main.741

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 11

58%

Researcher 7

37%

Lecturer / Post doc 1

5%

Readers' Discipline

Tooltip

Computer Science 19

83%

Engineering 2

9%

Neuroscience 1

4%

Medicine and Dentistry 1

4%

Save time finding and organizing research with Mendeley

Sign up for free