Multimodal named entity recognition (MNER) requires to bridge the gap between language understanding and visual context. While many multimodal neural techniques have been proposed to incorporate images into the MNER task, the model’s ability to leverage multimodal interactions remains poorly understood. In this work, we conduct in-depth analyses of existing multimodal fusion techniques from different perspectives and describe the scenarios where adding information from the image does not always boost performance. We also study the use of captions as a way to enrich the context for MNER. Experiments on three datasets from popular social platforms expose the bottleneck of existing multimodal models and the situations where using captions is beneficial.
CITATION STYLE
Chen, S., Aguilar, G., Neves, L., & Solorio, T. (2021). Can images help recognize entities? A study of the role of images for Multimodal NER. In W-NUT 2021 - 7th Workshop on Noisy User-Generated Text, Proceedings of the Conference (pp. 87–96). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2021.wnut-1.11
Mendeley helps you to discover research relevant for your work.