Study of using hybrid deep neural networks in character extraction from images containing text

  • P P
  • HR M
  • Hrishikesh V
N/ACitations
Citations of this article
5Readers
Mendeley users who have this article in their library.

Abstract

Character segmentation from epigraphical images helps the optical character recognizer (OCR) in training and recognition of old regional scripts. The scripts or characters present in the images are illegible and may have complex and noisy background texture. In this paper, we present an automated way of segmenting and extracting characters on digitized inscriptions. To achieve this, machine learning models are employed to discern between correctly segmented characters and partially segmented ones. The proposed method first recursively crops the document by sliding a window across the image from top to bottom to extract the content within the window. This results in a number of small images for classification. The segments are classified into character and non-character class based on the features within them. The model was tested on a wide range of input images having irregular, inconsistently spaced, hand written and inscribed characters.

Cite

CITATION STYLE

APA

P, P., HR, M., & Hrishikesh, V. (2021). Study of using hybrid deep neural networks in character extraction from images containing text. Trends in Computer Science and Information Technology, 045–052. https://doi.org/10.17352/tcsit.000039

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free