Teaching Yourself: A Self-Knowledge Distillation Approach to Action Recognition

33Citations
Citations of this article
31Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

Knowledge distillation, which is a process of transferring complex knowledge learned by a heavy network, i.e., a teacher, to a lightweight network, i.e., a student, has emerged as an effective technique for compressing neural networks. To reduce the necessity of training a large teacher network, this paper leverages the recent self-knowledge distillation approach to train a student network progressively by distilling its own knowledge without a pre-trained teacher network. Far from the existing self-knowledge distillation methods, which mainly focus on still images, our proposed Teaching Yourself is a self-knowledge distillation technique that targets at videos for human action recognition. Our proposed Teaching Yourself is not only designed as an effective lightweight network but also a high generalization capability model. In our approach, the network is able to update itself using the best past model, termed the preceding model, which is then utilized to guide the training process to update the present model. Inspired by consistency training in state-of-the-art semi-supervised learning methods, we also introduce an effective augmentation strategy to increase data diversity and improve network generalization and consistent predictions for our proposed Teaching Yourself approach. Our benchmark has been conducted on both the 3D Resnet-18 and 3D ResNet-50 backbone networks and evaluated on various standard datasets such as UCF101, HMDB51, and Kinetics400 datasets. The experimental results have shown that our teaching yourself method significantly improves the action recognition performance in terms of accuracy compared to existing supervised learning and knowledge distillation methods. We also have conducted an expensive ablation study to demonstrate that our approach mitigates overconfident predictions on dark knowledge and generates more consistent predictions in input variations of the same data point. The code is available at https://github.com/vdquang1991/Self-KD.

References Powered by Scopus

Deep residual learning for image recognition

174876Citations
N/AReaders
Get full text

Squeeze-and-Excitation Networks

26088Citations
N/AReaders
Get full text

Learning spatiotemporal features with 3D convolutional networks

7893Citations
N/AReaders
Get full text

Cited by Powered by Scopus

A Survey on Efficient Convolutional Neural Networks and Hardware Acceleration

128Citations
N/AReaders
Get full text

CLIP-TSA: Clip-Assisted Temporal Self-Attention for Weakly-Supervised Video Anomaly Detection

30Citations
N/AReaders
Get full text

(2+1)D Distilled ShuffleNet: A Lightweight Unsupervised Distillation Network for Human Action Recognition

13Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Vu, D. Q., Le, N., & Wang, J. C. (2021). Teaching Yourself: A Self-Knowledge Distillation Approach to Action Recognition. IEEE Access, 9, 105711–105723. https://doi.org/10.1109/ACCESS.2021.3099856

Readers' Seniority

Tooltip

Lecturer / Post doc 3

33%

PhD / Post grad / Masters / Doc 3

33%

Researcher 3

33%

Readers' Discipline

Tooltip

Computer Science 6

55%

Engineering 3

27%

Chemistry 1

9%

Social Sciences 1

9%

Save time finding and organizing research with Mendeley

Sign up for free