STELA: A Real-Time Scene Text Detector with Learned Anchor

23Citations
Citations of this article
45Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

To achieve high coverage of target boxes, a normal strategy of conventional one-stage anchor-based detectors is to utilize multiple priors at each spatial position, especially in scene text detection tasks. In this work, we present a simple and intuitive method for multi-oriented text detection where each location of feature maps only associates with one reference box. The idea is inspired from the two-stage R-CNN framework that can estimate the location of objects with any shape by using learned proposals. The aim of our method is to integrate this mechanism into a one-stage detector and employ the learned anchor which is obtained through a regression operation to replace the original one into the final predictions. Based on RetinaNet, our method achieves competitive performances on several public benchmarks with a totally real-time efficiency (26.5fps at 800p), which surpasses all of anchor-based scene text detectors. In addition, with less attention on anchor design, we believe our method is easy to be applied on other analogous detection tasks. The code is publicly available at https://github.com/xhzdeng/stela.

Cite

CITATION STYLE

APA

Deng, L., Gong, Y., Lu, X., Lin, Y., Ma, Z., & Xie, M. (2019). STELA: A Real-Time Scene Text Detector with Learned Anchor. IEEE Access, 7, 153400–153407. https://doi.org/10.1109/ACCESS.2019.2948405

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free