Vision Transformers (ViTs) have been widely used in large-scale Vision and Language Pretraining (VLP) models. Though previous VLP works have proved the effectiveness of ViTs, they still suffer from computational efficiency brought by the long visual sequence. To tackle this problem, in this paper, we propose an efficient vision-and-language pre-training model with Text-Relevant Image Patch Selection, namely TRIPS, which reduces the visual sequence progressively with a text-guided patch-selection layer in the visual backbone for efficient training and inference. The patch-selection layer can dynamically compute text-dependent visual attention to identify the attentive image tokens with text guidance and fuse inattentive ones in an end-to-end manner. Meanwhile, TRIPS does not introduce extra parameters to ViTs. Experimental results on a variety of popular benchmark datasets demonstrate that TRIPS gain a speedup of 40% over previous similar VLP models, yet with competitive or better downstream task performance.
CITATION STYLE
Jiang, C., Xu, H., Li, C., Yan, M., Ye, W., Zhang, S., … Huang, S. (2022). TRIPS: Efficient Vision-and-Language Pre-training with Text-Relevant Image Patch Selection. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, EMNLP 2022 (pp. 4084–4096). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2022.emnlp-main.273
Mendeley helps you to discover research relevant for your work.