There are 68 repositories under talking-head topic.
[CVPR 2023] SadTalker:Learning Realistic 3D Motion Coefficients for Stylized Audio-Driven Single Image Talking Face Animation
AudioGPT: Understanding and Generating Speech, Music, Sound, and Talking Head
Real time interactive streaming digital human
[CVPR 2022] Thin-Plate Spline Motion Model for Image Animation.
EchoMimicV2: Towards Striking, Simplified, and Semi-Body Human Animation
[CVPR 2024] This is the official source for our paper "SyncTalk: The Devil is in the Synchronization for Talking Head Synthesis"
Wav2Lip UHQ extension for Automatic1111
Official code for CVPR2022 paper: Depth-Aware Generative Adversarial Network for Talking Head Video Generation
💬 An extensive collection of exceptional resources dedicated to the captivating world of talking face synthesis! ⭐ If you find this repo useful, please give it a star! 🤩
CVPR2023 talking face implementation for Identity-Preserving Talking Face Generation With Landmark and Appearance Priors
实时语音交互数字人,支持端到端语音方案(GLM-4-Voice - THG)和级联方案(ASR-LLM-TTS-THG)。可自定义形象与音色,无须训练,支持音色克隆,首包延迟低至3s。Real-time voice interactive digital human, supporting end-to-end voice solutions (GLM-4-Voice - THG) and cascaded solutions (ASR-LLM-TTS-THG). Customizable appearance and voice, supporting voice cloning, with initial package delay as low as 3s.
ICASSP 2022: "Text2Video: text-driven talking-head video synthesis with phonetic dictionary".
Talking Head (3D): A JavaScript class for real-time lip-sync using Ready Player Me full-body 3D avatars.
[ECCV 2024 Oral] EDTalk - Official PyTorch Implementation
A Survey on Deepfake Generation and Detection
code for paper "Audio2Head: Audio-driven One-shot Talking-head Generation with Natural Head Motion" in the conference of IJCAI 2021
The official code of our ICCV2023 work: Implicit Identity Representation Conditioned Memory Compensation Network for Talking Head video Generation
Long-Inference, High Quality Synthetic Speaker (AI avatar/ AI presenter)
DiffPoseTalk: Speech-Driven Stylistic 3D Facial Animation and Head Pose Generation via Diffusion Models
[CVPR 2024] FaceTalk: Audio-Driven Motion Diffusion for Neural Parametric Head Models
Freeform Body Motion Generation from Speech
Avatar Generation For Characters and Game Assets Using Deep Fakes
Offical implement of Dynamic Frame Avatar with Non-autoregressive Diffusion Framework for talking head Video Generation
The authors' implementation of the "Neural Head Reenactment with Latent Pose Descriptors" (CVPR 2020) paper.
The pytorch implementation of our WACV23 paper "Cross-identity Video Motion Retargeting with Joint Transformation and Synthesis".
Code for ACCV 2020 "Speech2Video Synthesis with 3D Skeleton Regularization and Expressive Body Poses"
Crystal TTVS engine is a real-time audio-visual Multilingual speech synthesizer with a 3D expressive avatar.
One-shot Audio-driven 3D Talking Head Synthesis via Generative Prior, CVPRW 2024