A Historical Survey of Advances in Transformer Architectures

1Citations
Citations of this article
20Readers
Mendeley users who have this article in their library.

Abstract

In recent times, transformer-based deep learning models have risen in prominence in the field of machine learning for a variety of tasks such as computer vision and text generation. Given this increased interest, a historical outlook at the development and rapid progression of transformer-based models becomes imperative in order to gain an understanding of the rise of this key architecture. This paper presents a survey of key works related to the early development and implementation of transformer models in various domains such as generative deep learning and as backbones of large language models. Previous works are classified based on their historical approaches, followed by key works in the domain of text-based applications, image-based applications, and miscellaneous applications. A quantitative and qualitative analysis of the various approaches is presented. Additionally, recent directions of transformer-related research such as those in the biomedical and timeseries domains are discussed. Finally, future research opportunities, especially regarding the multi-modality and optimization of the transformer training process, are identified.

Cite

CITATION STYLE

APA

Sajun, A. R., Zualkernan, I., & Sankalpa, D. (2024, May 1). A Historical Survey of Advances in Transformer Architectures. Applied Sciences (Switzerland). Multidisciplinary Digital Publishing Institute (MDPI). https://doi.org/10.3390/app14104316

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free