Improving Intra- And Inter-Modality Visual Relation for Image Captioning

28Citations
Citations of this article
24Readers
Mendeley users who have this article in their library.
Get full text

Abstract

It is widely shared that capturing relationships among multi-modality features would be helpful for representing and ultimately describing an image. In this paper, we present a novel Intra- and Inter-modality visual Relation Transformer to improve connections among visual features, termed I2RT. Firstly, we propose Relation Enhanced Transformer Block (RETB) for image feature learning, which strengthens intra-modality visual relations among objects. Moreover, to bridge the gap between inter-modality feature representations, we align them explicitly via Visual Guided Alignment (VGA) module. Finally, an end-to-end formulation is adopted to train the whole model jointly. Experiments on the MS-COCO dataset show the effectiveness of our model, leading to improvements on all commonly used metrics on the "Karpathy"test split. Extensive ablation experiments are conducted for the comprehensive analysis of the proposed method.

Cite

CITATION STYLE

APA

Wang, Y., Zhang, W. K., Liu, Q., Zhang, Z., Gao, X., & Sun, X. (2020). Improving Intra- And Inter-Modality Visual Relation for Image Captioning. In MM 2020 - Proceedings of the 28th ACM International Conference on Multimedia (pp. 4190–4198). Association for Computing Machinery, Inc. https://doi.org/10.1145/3394171.3413877

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free