Show, Attend and Distill: Knowledge Distillation via Attention-based Feature Matching

102Citations
Citations of this article
118Readers
Mendeley users who have this article in their library.

Abstract

Knowledge distillation extracts general knowledge from a pre-trained teacher network and provides guidance to a target student network. Most studies manually tie intermediate features of the teacher and student, and transfer knowledge through pre-defined links. However, manual selection often constructs ineffective links that limit the improvement from the distillation. There has been an attempt to address the problem, but it is still challenging to identify effective links under practical scenarios. In this paper, we introduce an effective and efficient feature distillation method utilizing all the feature levels of the teacher without manually selecting the links. Specifically, our method utilizes an attention-based meta-network that learns relative similarities between features, and applies identified similarities to control distillation intensities of all possible pairs. As a result, our method determines competent links more efficiently than the previous approach and provides better performance on model compression and transfer learning tasks. Further qualitative analyses and ablative studies describe how our method contributes to better distillation. The implementation code is available at open sourced.

Cite

CITATION STYLE

APA

Ji, M., Heo, B., & Park, S. (2021). Show, Attend and Distill: Knowledge Distillation via Attention-based Feature Matching. In 35th AAAI Conference on Artificial Intelligence, AAAI 2021 (Vol. 9B, pp. 7945–7952). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v35i9.16969

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free