E2E-MLT - An Unconstrained End-to-End Method for Multi-language Scene Text

35Citations
Citations of this article
72Readers
Mendeley users who have this article in their library.
Get full text

Abstract

An end-to-end trainable (fully differentiable) method for multi-language scene text localization and recognition is proposed. The approach is based on a single fully convolutional network (FCN) with shared layers for both tasks. E2E-MLT is the first published multi-language OCR for scene text. While trained in multi-language setup, E2E-MLT demonstrates competitive performance when compared to other methods trained for English scene text alone. The experiments show that obtaining accurate multi-language multi-script annotations is a challenging problem. Code and trained models are released publicly at https://github.com/MichalBusta/E2E-MLT.

Cite

CITATION STYLE

APA

Bušta, M., Patel, Y., & Matas, J. (2019). E2E-MLT - An Unconstrained End-to-End Method for Multi-language Scene Text. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 11367 LNCS, pp. 127–143). Springer Verlag. https://doi.org/10.1007/978-3-030-21074-8_11

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free