There are 62 repositories under video-generation topic.
Bring portraits to life!
VideoCrafter2: Overcoming Data Limitations for High-Quality Video Diffusion Models
[ICCV 2023 Oral] Text-to-Image Diffusion Models are Zero-Shot Video Generators
Champ: Controllable and Consistent Human Image Animation with 3D Parametric Guidance
InternGPT (iGPT) is an open source demo platform where you can easily showcase your AI models. Now it supports DragGAN, ChatGPT, ImageBind, multimodal chat like GPT-4, SAM, interactive image editing, etc. Try it at igpt.opengvlab.com (支持DragGAN、ChatGPT、ImageBind、SAM的在线Demo系统)
A curated list of recent diffusion models for video generation, editing, restoration, understanding, etc.
MuseV: Infinite-length and High Fidelity Virtual Human Video Generation with Visual Conditioned Parallel Denoising
Official implementations for paper: DreamTalk: When Expressive Talking Head Generation Meets Diffusion Probabilistic Models
MagicTime: Time-lapse Video Generation Models as Metamorphic Simulators
High-Quality Human Motion Video Generation with Confidence-aware Pose Guidance
Code for Motion Representations for Articulated Animation paper
Implementation of Video Diffusion Models, Jonathan Ho's new paper extending DDPMs to Video Generation - in Pytorch
[AAAI 2024] Follow-Your-Pose: This repo is the official implementation of "Follow-Your-Pose : Pose-Guided Text-to-Video Generation using Pose-Free Videos"
MiniSora: A community aims to explore the implementation path and future development direction of Sora.
Official JAX implementation of MAGVIT: Masked Generative Video Transformer
Official repo for VideoComposer: Compositional Video Synthesis with Motion Controllability
[ICLR 2024] SEINE: Short-to-Long Video Diffusion Model for Generative Transition and Prediction
[arXiv 2024] Follow-Your-Click: This repo is the official implementation of "Follow-Your-Click: Open-domain Regional Image Animation via Short Prompts"
MotionDirector: Motion Customization of Text-to-Video Diffusion Models.
[ICLR 2024] Official pytorch implementation of "ControlVideo: Training-free Controllable Text-to-Video Generation"
A system of bots that collects clips automatically via custom made filters, lets you easily browse these clips, and puts them together into a compilation video ready to be uploaded straight to any social media platform. Full VPS support is provided, along with an accounts system so multiple users can use the bot at once. This bot is split up into three separate programs. The server. The client. The video generator. These programs perform different functions that when combined creates a very powerful system for auto generating compilation videos.
Fine-Grained Open Domain Image Animation with Motion Guidance
[ICML 2024] MagicPose(also known as MagicDance): Realistic Human Poses and Facial Expressions Retargeting with Identity-aware Diffusion
Code for Paper "UniAnimate: Taming Unified Video Diffusion Models for Consistent Human Image Animation".
AnimateLCM: Let's Accelerate the Video Generation within 4 Steps!
A Survey on Text-to-Video Generation/Synthesis.
Implementation of MagViT2 Tokenizer in Pytorch
Multimodal AI Story Teller, built with Stable Diffusion, GPT, and neural text-to-speech
[ECCV 2024] FreeInit: Bridging Initialization Gap in Video Diffusion Models
[ICLR24] Official implementation of the paper “MagicDrive: Street View Generation with Diverse 3D Geometry Control”
The pytorch implementation of our CVPR 2023 paper "Conditional Image-to-Video Generation with Latent Flow Diffusion Models"
A Generalizable World Model for Autonomous Driving
A Collection of Papers and Codes for CVPR2024/ECCV2024 AIGC
[CVPR'23] MM-Diffusion: Learning Multi-Modal Diffusion Models for Joint Audio and Video Generation