Improving shape deformation in unsupervised image-to-image translation

6Citations
Citations of this article
262Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

Unsupervised image-to-image translation techniques are able to map local texture between two domains, but they are typically unsuccessful when the domains require larger shape change. Inspired by semantic segmentation, we introduce a discriminator with dilated convolutions that is able to use information from across the entire image to train a more context-aware generator. This is coupled with a multi-scale perceptual loss that is better able to represent error in the underlying shape of objects. We demonstrate that this design is more capable of representing shape deformation in a challenging toy dataset, plus in complex mappings with significant dataset variation between humans, dolls, and anime faces, and between cats and dogs.

Cite

CITATION STYLE

APA

Gokaslan, A., Ramanujan, V., Ritchie, D., Kim, K. I., & Tompkin, J. (2018). Improving shape deformation in unsupervised image-to-image translation. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 11216 LNCS, pp. 662–678). Springer Verlag. https://doi.org/10.1007/978-3-030-01258-8_40

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free