Understanding advertisements with BERT

5Citations
Citations of this article
86Readers
Mendeley users who have this article in their library.

Abstract

We consider a task based on CVPR 2018 challenge dataset on advertisement (Ad) understanding. The task involves detecting the viewer's interpretation of an Ad image captured as text. Recent results have shown that the embedded scene-text in the image holds a vital cue for this task. Motivated by this, we fine-tune the base BERT model for a sentence-pair classification task. Despite utilizing the scene-text as the only source of visual information, we could achieve a hit-or-miss accuracy of 84.95% on the challenge test data. To enable BERT to process other visual information, we append image captions to the scene-text. This achieves an accuracy of 89.69%, which is an improvement of 4.7%. This is the best reported result for this task.

Cite

CITATION STYLE

APA

Kalra, K., Kurma, B., Sreelatha, S., Patwardhan, M., & Karande, S. (2020). Understanding advertisements with BERT. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 7542–7547). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2020.acl-main.674

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free