Abstract
Vulnerability to lexical perturbation is a critical weakness of automatic evaluation metrics for image captioning. This paper proposes Perturbation Robust Multi-Lingual CLIPScore(PR-MCS), which exhibits robustness to such perturbations, as a novel reference-free image captioning metric applicable to multiple languages. To achieve perturbation robustness, we fine-tune the text encoder of CLIP with our language-agnostic method to distinguish the perturbed text from the original text. To verify the robustness of PR-MCS, we introduce a new fine-grained evaluation dataset consisting of detailed captions, critical objects, and the relationships between the objects for 3, 000 images in five languages. In our experiments, PR-MCS significantly outperforms baseline metrics in capturing lexical noise of all various perturbation types in all five languages, while maintaining a strong correlation with human judgments.
Cite
CITATION STYLE
Kim, Y., Hwang, Y., Yun, H., Yoon, S., Bui, T., & Jung, K. (2023). PR-MCS: Perturbation Robust Metric for MultiLingual Image Captioning. In Findings of the Association for Computational Linguistics: EMNLP 2023 (pp. 12237–12258). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-emnlp.819
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.