Abstract
While progress has been made in the field of portrait reenactment, the problem of how to produce high-fidelity and robust videos remains. Recent studies normally find it challenging to handle rarely seen target poses due to the limitation of source data. This paper proposes the Video Portrait via Non-local Quantization Modeling (VPNQ) framework, which produces pose- and disturbance-robust reenactable video portraits. Our key insight is to learn position-invariant quantized local patch representations, then build a mapping between simple driving signals and local textures with non-local spatial-temporal modeling. Specifically, instead of learning a universal quantized codebook, we identify that a personalized one can be trained to preserve desired position-invariant local details. Then, a simple representation of projected landmarks can be used as sufficient driving signals to avoid 3D rendering. In the following, we employ a carefully designed Spatio-Temporal Transformer to predict reasonable and temporally consistent quantized tokens from the driving signal. The predicted codes can be decoded back to robust and high-quality videos. Comprehensive experiments have been conducted to validate the effectiveness of our approach.
Cite
CITATION STYLE
Wang, K., Liang, C., Zhou, H., Tang, J., Wu, Q., He, D., … Wang, J. (2023). Robust Video Portrait Reenactment via Personalized Representation Quantization. In Proceedings of the 37th AAAI Conference on Artificial Intelligence, AAAI 2023 (Vol. 37, pp. 2564–2572). AAAI Press. https://doi.org/10.1609/aaai.v37i2.25354
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.