CapNet: An Encoder-Decoder based Neural Network Model for Automatic Bangla Image Caption Generation

1Citations
Citations of this article
8Readers
Mendeley users who have this article in their library.

Abstract

Automatic caption generation from images has become an active research topic in the field of Computer Vision (CV) and Natural Language Processing (NLP). Machine generated image caption plays a vital role for the visually impaired people by converting the caption to speech to have a better understanding of their surrounding. Though significant amount of research has been conducted for automatic caption generation in other languages, far too little effort has been devoted to Bangla image caption generation. In this paper, we propose an encoder-decoder based model which takes an image as input and generates the corresponding Bangla caption as output. The encoder network consists of a pretrained image feature extractor called ResNet-50, while the decoder network consists of Bidirectional LSTMs for caption generation. The model has been trained and evaluated using a Bangla image captioning dataset named BanglaLekhaImageCaptions. The proposed model achieved a training accuracy of 91% and BLEU-1, BLEU-2, BLEU-3, BLEU-4 scores of 0.81, 0.67, 0.57, and 0.51 respectively. Moreover, a comparative study for different pretrained feature extractors such as VGG-16 and Xception is presented. Finally, the proposed model has been deployed on an embedded device for analysing the inference time and power consumption.

Cite

CITATION STYLE

APA

Rahman, R., Murad, H., Rahman, N. N., Saha, A. K., Al Masud, S. M. R., & Momtaz, A. S. Z. (2022). CapNet: An Encoder-Decoder based Neural Network Model for Automatic Bangla Image Caption Generation. International Journal of Advanced Computer Science and Applications, 13(8), 752–759. https://doi.org/10.14569/IJACSA.2022.0130886

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free