There are 2 repositories under controllable-generation topic.
PyTorch implementation for Score-Based Generative Modeling through Stochastic Differential Equations (ICLR 2021, Oral)
Caption-Anything is a versatile tool combining image segmentation, visual captioning, and ChatGPT, generating tailored captions with diverse controls for user preferences. https://huggingface.co/spaces/TencentARC/Caption-Anything https://huggingface.co/spaces/VIPLab/Caption-Anything
Official code for Score-Based Generative Modeling through Stochastic Differential Equations (ICLR 2021, Oral)
PyTorch implementation for SDEdit: Image Synthesis and Editing with Stochastic Differential Equations
A collection of resources on controllable generation with text-to-image diffusion models.
[ECCV 2024] MOFA-Video: Controllable Image Animation via Generative Motion Field Adaptions in Frozen Image-to-Video Diffusion Model.
StyleShot: A SnapShot on Any Style. 一款可以迁移任意风格到任意内容的模型,无需针对图片微调,即能生成高质量的个性风格化图片!
Official implementation of "DiffuseVAE: Efficient, Controllable and High-Fidelity Generation from Low-Dimensional Latents"
Update-to-data resources for conditional content generation, including human motion generation, image or video generation and editing.
[CVPR 2024] Official implementation of "Towards Realistic Scene Generation with LiDAR Diffusion Models"
[ICLR 2025] Codebase for "CtrLoRA: An Extensible and Efficient Framework for Controllable Image Generation"
[3DV-2025] Official implementation of "Controllable Text-to-3D Generation via Surface-Aligned Gaussian Splatting"
A curated list of resources of audio-driven talking face generation
[ECCV 2024] AnyControl, a multi-control image synthesis model that supports any combination of user provided control signals. 一个支持用户自由输入控制信号的图像生成模型,能够根据多种控制生成自然和谐的结果!
[ICML 2023] Official PyTorch Implementation of "Hierarchical Neural Coding for Controllable CAD Model Generation".
Official PyTorch Implementation for InfoSwap
[Paperlist] Awesome paper list of controllable text generation via latent auto-encoders. Contributions of any kind are welcome.
This is the project for 'Any2Caption', Interpreting Any Condition to Caption for Controllable Video Generation
awesome-LLM-controlled-constrained-generation
PerLDiff: Controllable Street View Synthesis Using Perspective-Layout Diffusion Models
[Preprint] AdaVAE: Exploring Adaptive GPT-2s in VAEs for Language Modeling PyTorch Implementation
Implementation of Collage Diffusion (https://arxiv.org/abs/2303.00262)
[KBS] PCAE: A Framework of Plug-in Conditional Auto-Encoder for Controllable Text Generation PyTorch Implementation
Controllable mage captioning model with unsupervised modes
Awesome Controllable Video Generation with Diffusion Models
Using diffusion model to reach controllable end-to-end driving with Carla simulation environment.
Source code of our TACL paper "Controllable Summarization with Constrained Markov Decision Process"
Code for SAPPHIRE: Approaches for Enhanced Concept-to-Text Generation (https://aclanthology.org/2021.inlg-1.21/) INLG 2021 Best Long Paper.
Unsupervised Discovery of Steerable Factors When Graph Deep Generative Models Are Entangled, TMLR 2024
Official repo for the paper "Mojito: Motion Trajectory and Intensity Control for Video Generation""
TTIDA: Controllable Generative Data Augmentation via Text-to-Text and Text-to-Image Models