Transformer-Exclusive Cross-Modal Representation for Vision and Language

2Citations
Citations of this article
47Readers
Mendeley users who have this article in their library.

Abstract

Ever since the advent of deep learning, cross-modal representation learning has been dominated by the approaches involving convolutional neural networks for visual representation and recurrent neural networks for language representation. Transformer architecture, however, has rapidly taken over the recurrent neural networks in natural language processing tasks, and it has also been shown that vision tasks can be handled with transformer architecture, with compatible performance to convolutional neural networks. Such results naturally lead to speculation upon the possibility of tackling cross-modal representation for vision and language exclusively with transformer. This paper examines transformer-exclusive cross-modal representation to explore such possibility, demonstrating its potentials as well as discussing its current limitations and its prospects.

Cite

CITATION STYLE

APA

Shin, A., & Narihira, T. (2021). Transformer-Exclusive Cross-Modal Representation for Vision and Language. In Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021 (pp. 2719–2725). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.findings-acl.240

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free