DisCoHead: Audio-and-Video-Driven Talking Head Generation by Disentangled Control of Head Pose and Facial Expressions
You can install required environments using below commands:
git clone https://github.com/deepbrainai-research/discohead
cd discohead
conda create -n discohead python=3.7
conda activate discohead
conda install pytorch==1.10.0 torchvision==0.11.1 torchaudio==0.10.0 cudatoolkit=10.2 -c pytorch
pip install -r requirements.txt
- Download the pre-trained checkpoints from google drive and put into
weight
folder. - Download
dataset.zip
from google drive and unzip intodataset
. DisCoHead
directory should have the following structure.
DisCoHead/
├── dataset/
│ ├── grid/
│ │ ├── demo1/
│ │ ├── demo2/
│ ├── koeba/
│ │ ├── demo1/
│ │ ├── demo2/
│ ├── obama/
│ │ ├── demo1/
│ │ ├── demo2/
├── weight/
│ ├── grid.pt
│ ├── koeba.pt
│ ├── obama.pt
├── modules/
‥‥
- The
--mode
argument is used to specify which demo video you want to generate:
python test.py --mode {mode}
- Available modes:
obama_demo1, obama_demo2, grid_demo1, grid_demo2, koeba_demo1, koeba_demo2
Non-commercial
To be updated