Distilling Distribution Knowledge in Normalizing Flow

1Citations
Citations of this article
2Readers
Mendeley users who have this article in their library.

Abstract

In this letter, we propose a feature-based knowledge distillation scheme which transfers knowledge between intermediate blocks of teacher and student with flow-based architecture, specifically Normalizing flow in our implementation. In addition to the knowledge transfer scheme, we examine how configuration of the distillation positions impacts on the knowledge transfer performance. To evaluate the proposed ideas, we choose two knowledge distillation baseline models which are based on Normalizing flow on different domains: CS-Flow for anomaly detection and SRFlow-DA for super-resolution. A set of performance comparison to the baseline models with popular benchmark datasets shows promising results along with improved inference speed. The comparison includes performance analysis based on various configurations of the distillation positions in the proposed scheme.

Cite

CITATION STYLE

APA

Kwon, J., & Kim, G. (2023). Distilling Distribution Knowledge in Normalizing Flow. IEICE Transactions on Information and Systems, E106.D(8), 1287–1291. https://doi.org/10.1587/transinf.2022EDL8103

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free