Fully Supervised and Guided Distillation for One-Stage Detectors

1Citations
Citations of this article
16Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Model distillation has been extended from image classification to object detection. However, existing approaches are difficult to focus on both object regions and false detection regions of student networks to effectively distill the feature representation from teacher networks. To address it, we propose a fully supervised and guided distillation algorithm for one-stage detectors, where an excitation and suppression loss is designed to make a student network mimic the feature representation of a teacher network in the object regions and its own high-response regions in the background, so as to excite the feature expression of object regions and adaptively suppress the feature expression of high-response regions that may cause false detections. Besides, a process-guided learning strategy is proposed to train the teacher along with the student and transfer knowledge throughout the training process. Extensive experiments on Pascal VOC and COCO benchmarks demonstrate the following advantages of our algorithm, including the effectiveness for improving recall and reducing false detections, the robustness on common one-stage detector heads and the superiority compared with state-of-the-art methods.

Cite

CITATION STYLE

APA

Wang, D., Wen, D., Liu, J., Tao, W., Chen, T. W., Osa, K., & Kato, M. (2021). Fully Supervised and Guided Distillation for One-Stage Detectors. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 12624 LNCS, pp. 171–188). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-030-69535-4_11

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free