MOST-GAN: 3D Morphable StyleGAN for Disentangled Face Image Manipulation

21Citations
Citations of this article
49Readers
Mendeley users who have this article in their library.

Abstract

Recent advances in generative adversarial networks (GANs) have led to remarkable achievements in face image synthesis. While methods that use style-based GANs can generate strikingly photorealistic face images, it is often difficult to control the characteristics of the generated faces in a meaningful and disentangled way. Prior approaches aim to achieve such semantic control and disentanglement within the latent space of a previously trained GAN. In contrast, we propose a framework that a priori models physical attributes of the face such as 3D shape, albedo, pose, and lighting explicitly, thus providing disentanglement by design. Our method, MOST-GAN, integrates the expressive power and photorealism of style-based GANs with the physical disentanglement and flexibility of nonlinear 3D morphable models, which we couple with a state-of-the-art 2D hair manipulation network. MOST-GAN achieves photorealistic manipulation of portrait images with fully disentangled 3D control over their physical attributes, enabling extreme manipulation of lighting, facial expression, and pose variations up to full profile view.

References Powered by Scopus

U-net: Convolutional networks for biomedical image segmentation

64903Citations
N/AReaders
Get full text

Perceptual losses for real-time style transfer and super-resolution

7397Citations
N/AReaders
Get full text

A style-based generator architecture for generative adversarial networks

6822Citations
N/AReaders
Get full text

Cited by Powered by Scopus

DCFace: Synthetic Face Generation with Dual Condition Diffusion Model

65Citations
N/AReaders
Get full text

Proactive Image Manipulation Detection

26Citations
N/AReaders
Get full text

GAN-Based Facial Attribute Manipulation

24Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Medin, S. C., Egger, B., Cherian, A., Wang, Y., Tenenbaum, J. B., Liu, X., & Marks, T. K. (2022). MOST-GAN: 3D Morphable StyleGAN for Disentangled Face Image Manipulation. In Proceedings of the 36th AAAI Conference on Artificial Intelligence, AAAI 2022 (Vol. 36, pp. 1962–1971). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v36i2.20091

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 16

84%

Researcher 2

11%

Lecturer / Post doc 1

5%

Readers' Discipline

Tooltip

Computer Science 18

86%

Engineering 2

10%

Agricultural and Biological Sciences 1

5%

Save time finding and organizing research with Mendeley

Sign up for free