Camera captured DIQA with linearity and monotonicity constraints

5Citations
Citations of this article
7Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Document image quality assessment (DIQA), which predicts the visual quality of the document images, can not only be applied to estimate document’s optical character recognition (OCR) performance prior to any actual recognition, but also provides immediate feedback on whether the documents meet the quality requirements for other high level document processing and analysis tasks. In this work, we present a deep neural network (DNN) to accomplish the DIQA task, where a Saimese based deep convolutional neural network (DCNN) is employed with customized losses to improve system’s capability of linearity and monotonicity to predict the quality of document images. Based on the proposed network along with the new losses, the obtained DCNN achieves the state-of-the-art quality assessment performance on the public datasets. The source codes and pre-trained models are available at https://gitlab.com/xujun.peng/DIQA-linearity-monotonicity.

Cite

CITATION STYLE

APA

Peng, X., & Wang, C. (2020). Camera captured DIQA with linearity and monotonicity constraints. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 12116 LNCS, pp. 168–181). Springer. https://doi.org/10.1007/978-3-030-57058-3_13

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free