Neural caption generation over figures

17Citations
Citations of this article
11Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Figures are human-friendly but difficult for computers to process automatically. In this work, we investigate the problem of figure captioning. The goal is to automatically generate a natural language description of a given figure. We create a new dataset for figure captioning, FigCAP. To achieve accurate generation of labels in figures, we propose the Label Maps Attention Model. Extensive experiments show that our method outperforms the baselines. A successful solution to this task allows figure content to be accessible to those with visual impairment by providing input to a text-to-speech system; and enables automatic parsing of vast repositories of documents where figures are pervasive.

Cite

CITATION STYLE

APA

Chen, C., Zhang, R., Koh, E., Kim, S., Cohen, S., Yu, T., … Bunescu, R. (2019). Neural caption generation over figures. In UbiComp/ISWC 2019- - Adjunct Proceedings of the 2019 ACM International Joint Conference on Pervasive and Ubiquitous Computing and Proceedings of the 2019 ACM International Symposium on Wearable Computers (pp. 482–485). Association for Computing Machinery, Inc. https://doi.org/10.1145/3341162.3345601

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free