Efficient High-Resolution Image-to-Image Translation Using Multi-Scale Gradient U-Net

1Citations
Citations of this article
24Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Recently, Conditional Generative Adversarial Network (Conditional GAN) has shown very promising performance in several image-to-image translation applications. However, the uses of these conditional GANs are quite limited to low-resolution images, such as 256 × 256. The Pix2Pix-HD is a recent attempt to utilize the conditional GAN for high-resolution image synthesis. In this paper, we propose a Multi-Scale Gradient based U-Net (MSG U-Net) model for high-resolution image-to-image translation up to 2048 × 1024 resolution. The proposed model is trained by allowing the flow of gradients from multiple-discriminators to a single generator at multiple scales. The proposed MSG U-Net architecture leads to photo-realistic high-resolution image-to-image translation. Moreover, the proposed model is computationally efficient as compared to the Pix2Pix-HD with an improvement in the inference time nearly by 2.5 times. We provide the code of MSG U-Net model at https://github.com/laxmaniron/MSG-U-Net.

Cite

CITATION STYLE

APA

Laxman, K., Dubey, S. R., Kalyan, B., & Kojjarapu, S. R. V. (2022). Efficient High-Resolution Image-to-Image Translation Using Multi-Scale Gradient U-Net. In Communications in Computer and Information Science (Vol. 1567 CCIS, pp. 33–44). Springer Science and Business Media Deutschland GmbH. https://doi.org/10.1007/978-3-031-11346-8_4

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free