Distilling portable generative adversarial networks for image translation

69Citations
Citations of this article
74Readers
Mendeley users who have this article in their library.

Abstract

Despite Generative Adversarial Networks (GANs) have been widely used in various image-to-image translation tasks, they can be hardly applied on mobile devices due to their heavy computation and storage cost. Traditional network compression methods focus on visually recognition tasks, but never deal with generation tasks. Inspired by knowledge distillation, a student generator of fewer parameters is trained by inheriting the low-level and high-level information from the original heavy teacher generator. To promote the capability of student generator, we include a student discriminator to measure the distances between real images, and images generated by student and teacher generators. An adversarial learning process is therefore established to optimize student generator and student discriminator. Qualitative and quantitative analysis by conducting experiments on benchmark datasets demonstrate that the proposed method can learn portable generative models with strong performance.

Cite

CITATION STYLE

APA

Chen, H., Wang, Y., Shu, H., Wen, C., Xu, C., Shi, B., … Xu, C. (2020). Distilling portable generative adversarial networks for image translation. In AAAI 2020 - 34th AAAI Conference on Artificial Intelligence (pp. 3585–3592). AAAI press. https://doi.org/10.1609/aaai.v34i04.5765

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free