Identity-Aware and Shape-Aware Propagation of Face Editing in Videos | IEEE Journals & Magazine | IEEE Xplore

Identity-Aware and Shape-Aware Propagation of Face Editing in Videos


Abstract:

The development of deep generative models has inspired various facial image editing methods, but many of them are difficult to be directly applied to video editing due to...Show More

Abstract:

The development of deep generative models has inspired various facial image editing methods, but many of them are difficult to be directly applied to video editing due to various challenges ranging from imposing 3D constraints, preserving identity consistency, ensuring temporal coherence, etc. To address these challenges, we propose a new framework operating on the StyleGAN2 latent space for identity-aware and shape-aware edit propagation on face videos. In order to reduce the difficulties of maintaining the identity, keeping the original 3D motion, and avoiding shape distortions, we disentangle the StyleGAN2 latent vectors of human face video frames to decouple the appearance, shape, expression, and motion from identity. An edit encoding module is used to map a sequence of image frames to continuous latent codes with 3D parametric control and is trained in a self-supervised manner with identity loss and triple shape losses. Our model supports propagation of edits in various forms: I. direct appearance editing on a specific keyframe, II. implicit editing of face shape via a given reference image, and III. existing latent-based semantic edits. Experiments show that our method works well for various forms of videos in the wild and outperforms an animation-based approach and the recent deep generative techniques.
Page(s): 3444 - 3456
Date of Publication: 09 January 2023

ISSN Information:

PubMed ID: 37018564

Funding Agency:


Contact IEEE to Subscribe

References

References is not available for this document.