Abstract
Figures are human-friendly but difficult for computers to process automatically. In this work, we investigate the problem of figure captioning. The goal is to automatically generate a natural language description of a given figure. We create a new dataset for figure captioning, FigCAP. To achieve accurate generation of labels in figures, we propose the Label Maps Attention Model. Extensive experiments show that our method outperforms the baselines. A successful solution to this task allows figure content to be accessible to those with visual impairment by providing input to a text-to-speech system; and enables automatic parsing of vast repositories of documents where figures are pervasive.
Author supplied keywords
Cite
CITATION STYLE
Chen, C., Zhang, R., Koh, E., Kim, S., Cohen, S., Yu, T., … Bunescu, R. (2019). Neural caption generation over figures. In UbiComp/ISWC 2019- - Adjunct Proceedings of the 2019 ACM International Joint Conference on Pervasive and Ubiquitous Computing and Proceedings of the 2019 ACM International Symposium on Wearable Computers (pp. 482–485). Association for Computing Machinery, Inc. https://doi.org/10.1145/3341162.3345601
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.