In order to improve the results of automatically recognized handwritten text, information about the language is commonly included in the recognition process. A common approach is to represent a text line as a sequence. It is processed in one direction and the language information via n-grams is directly included in the decoding. This approach, however, only uses context on one side to estimate a word's probability. Therefore, we propose a bidirectional recognition in this paper, using distinct forward and a backward language models. By combining decoding hypotheses from both directions, we achieve a significant increase in recognition accuracy for the off-line writer independent handwriting recognition task. Both language models are of the same type and can be estimated on the same corpus. Hence, the increase in recognition accuracy comes without any additional need for training data or language modeling complexity. © 2012 Springer-Verlag Berlin Heidelberg.
CITATION STYLE
Frinken, V., Fornés, A., Lladós, J., & Ogier, J. M. (2012). Bidirectional language model for handwriting recognition. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 7626 LNCS, pp. 611–619). https://doi.org/10.1007/978-3-642-34166-3_67
Mendeley helps you to discover research relevant for your work.