A Shared Task on Multimodal Machine Translation and Crosslingual Image Description

187Citations
Citations of this article
140Readers
Mendeley users who have this article in their library.

Abstract

This paper introduces and summarises the findings of a new shared task at the intersection of Natural Language Processing and Computer Vision: the generation of image descriptions in a target language, given an image and/or one or more descriptions in a different (source) language. This challenge was organised along with the Conference on Machine Translation (WMT16), and called for system submissions for two task variants: (i) a translation task, in which a source language image description needs to be translated to a target language, (optionally) with additional cues from the corresponding image, and (ii) a description generation task, in which a target language description needs to be generated for an image, (optionally) with additional cues from source language descriptions of the same image. In this first edition of the shared task, 16 systems were submitted for the translation task and seven for the image description task, from a total of 10 teams.

Cite

CITATION STYLE

APA

Specia, L., Frank, S., Sima’An, K., & Elliott, D. (2016). A Shared Task on Multimodal Machine Translation and Crosslingual Image Description. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 2, pp. 543–553). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/w16-2346

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free