Image Caption Generation for News Articles

9Citations
Citations of this article
72Readers
Mendeley users who have this article in their library.

Abstract

In this paper, we address the task of news-image captioning, which generates a description of an image given the image and its article body as input. This task is more challenging than the conventional image captioning, because it requires a joint understanding of image and text. We present a Transformer model that integrates text and image modalities and attends to textual features from visual features in generating a caption. Experiments based on automatic evaluation metrics and human evaluation show that an article text provides primary information to reproduce news-image captions written by journalists. The results also demonstrate that the proposed model outperforms the state-of-the-art model. In addition, we also confirm that visual features contribute to improving the quality of news-image captions.

Cite

CITATION STYLE

APA

Yang, Z., & Okazaki, N. (2020). Image Caption Generation for News Articles. In COLING 2020 - 28th International Conference on Computational Linguistics, Proceedings of the Conference (pp. 1941–1951). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.coling-main.176

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free