LouisBearing / UnconditionalHeadMotion

Code & demo for the animation of still facial landmarks from an initial pose.

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Autoregressive GAN for Semantic Unconditional Head Motion Generation (SUHMo)

Abstract [Paper]

We address the task of unconditional head motion generation to animate still human faces in a low-dimensional semantic space. Deviating from talking head generation conditioned on audio that seldom puts emphasis on realistic head motions, we devise a GAN-based architecture that allows obtaining rich head motion sequences while avoiding known caveats associated with GANs. Namely, the autoregressive generation of incremental outputs ensures smooth trajectories, while a multi-scale discriminator on input pairs drives generation toward better handling of high and low frequency signals and less mode collapse. We demonstrate experimentally the relevance of the proposed architecture and compare with models that showed state-of-the-art performances on similar tasks.

Examplar results

In the results presented below 120 frames are generated from a single reference image.

SUHMo-RNN (Training on CONFER DB)

drawing drawing drawing drawing drawing drawing drawing drawing drawing drawing drawing drawing

SUHMo-Transformer (Training on VoxCeleb2)

~~ In Vox2 preprocessing faces are centered, hence the suppression of head translation ~~

drawing drawing drawing drawing drawing drawing drawing drawing drawing drawing drawing drawing

SUHMo in-the-wild

Several outputs can be obtained from the same reference image. See below for an illustration on SUHMo-RNN trained on CONFER DB.

drawing

drawingdrawingdrawingdrawingdrawingdrawing

drawingdrawingdrawingdrawingdrawingdrawing

drawingdrawingdrawingdrawingdrawingdrawing

drawingdrawingdrawingdrawingdrawingdrawing

Architecture overview

SUHMo is a framework that can be implemented in several forms. Below are the proposed LSTM and Transformer variants of our model.

uncond_head_mot

Execution & Pre-trained models

Incoming...

Citation

@misc{https://doi.org/10.48550/arxiv.2211.00987,
  doi = {10.48550/ARXIV.2211.00987},
  url = {https://arxiv.org/abs/2211.00987},
  author = {Airale, Louis and Alameda-Pineda, Xavier and Lathuilière, Stéphane and Vaufreydaz, Dominique},
  keywords = {Computer Vision and Pattern Recognition (cs.CV), FOS: Computer and information sciences, FOS: Computer and information sciences},
  title = {Autoregressive GAN for Semantic Unconditional Head Motion Generation},
  publisher = {arXiv},
  year = {2022},
  copyright = {arXiv.org perpetual, non-exclusive license}
}

References

Face Alignment

A. Bulat and G. Tzimiropoulos, “How far are we from solving the 2d & 3d face alignment problem? (and a dataset of 230,000 3d facial landmarks),” in ICCV, 2017.

CONFER DB

C. Georgakis, Y. Panagakis, S. Zafeiriou, and M. Pantic, “The conflict escalation resolution (confer) database,” Image and Vision Computing, vol. 65, 2017.

VoxCeleb2

J. S. Chung, A. Nagrani, and A. Zisserman, “Voxceleb2: Deep speaker recognition,” in INTERSPEECH, 2018.

About

Code & demo for the animation of still facial landmarks from an initial pose.

License:BSD 3-Clause "New" or "Revised" License


Languages

Language:Python 100.0%