Pose Manipulation with Identity Preservation

0Citations
Citations of this article
17Readers
Mendeley users who have this article in their library.

Abstract

This paper describes a new model which generates images in novel poses e.g. by altering face expression and orientation, from just a few instances of a human subject. Unlike previous approaches which require large datasets of a specific person for training, our approach may start from a scarce set of images, even from a single image. To this end, we introduce Character Adaptive Identity Normalization GAN (CainGAN) which uses spatial characteristic features extracted by an embedder and combined across source images. The identity information is propagated throughout the network by applying conditional normalization. After extensive adversarial training, CainGAN receives figures of faces from a certain individual and produces new ones while preserving the person's identity. Experimental results show that the quality of generated images scales with the size of the input set used during inference. Furthermore, quantitative measurements indicate that CainGAN performs better compared to other methods when training data is limited.

Cite

CITATION STYLE

APA

Ardelean, A. T., & Sasu, L. M. (2020). Pose Manipulation with Identity Preservation. International Journal of Computers, Communications and Control, 15(2), 1–9. https://doi.org/10.15837/IJCCC.2020.2.3862

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free