An end-to-end trainable (fully differentiable) method for multi-language scene text localization and recognition is proposed. The approach is based on a single fully convolutional network (FCN) with shared layers for both tasks. E2E-MLT is the first published multi-language OCR for scene text. While trained in multi-language setup, E2E-MLT demonstrates competitive performance when compared to other methods trained for English scene text alone. The experiments show that obtaining accurate multi-language multi-script annotations is a challenging problem. Code and trained models are released publicly at https://github.com/MichalBusta/E2E-MLT.
CITATION STYLE
Bušta, M., Patel, Y., & Matas, J. (2019). E2E-MLT - An Unconstrained End-to-End Method for Multi-language Scene Text. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 11367 LNCS, pp. 127–143). Springer Verlag. https://doi.org/10.1007/978-3-030-21074-8_11
Mendeley helps you to discover research relevant for your work.