There are 10 repositories under text-to-image-generation topic.
LTX-Video Support for ComfyUI
Custom Diffusion: Multi-Concept Customization of Text-to-Image Diffusion (CVPR 2023)
[CVPR 2025 Oral]Infinity ∞ : Scaling Bitwise AutoRegressive Modeling for High-Resolution Image Synthesis
[AAAI 2025]👔IMAGDressing👔: Interactive Modular Apparel Generation for Virtual Dressing. It enables customizable human image generation with flexible garment, pose, and scene control, ensuring high fidelity and garment consistency for virtual dressing.
Awesome Unified Multimodal Models
Rich-Text-to-Image Generation
UniWorld: High-Resolution Semantic Encoders for Unified Visual Understanding and Generation
(Accepted by IJCV) Liquid: Language Models are Scalable and Unified Multi-modal Generators
The most advanced Nano Banana image generator and editor application. Your central hub for AI image generation and revisions. Intuitive UI features reference images, editing with image masks, version history, and more. Powered by Gemini 2.5 Flash images API.
AutoStudio: Crafting Consistent Subjects in Multi-turn Interactive Image Generation
The Paper List of Large Multi-Modality Model (Perception, Generation, Unification), Parameter-Efficient Finetuning, Vision-Language Pretraining, Conventional Image-Text Matching for Preliminary Insight.
[CVPR 2025] 🔥 Official impl. of "TokenFlow: Unified Image Tokenizer for Multimodal Understanding and Generation".
[NeurIPS'23] "MagicBrush: A Manually Annotated Dataset for Instruction-Guided Image Editing".
Colab notebook for Stable Diffusion Hyper-SDXL.
[CVPR 2025] Aesthetic Post-Training Diffusion Models from Generic Preferences with Step-by-step Preference Optimization
Official repository for "CFG++: manifold-constrained classifier free guidance for diffusion models" (ICLR2025)
Faster generation with text-to-image diffusion models.
[NeurIPS-2023] Annual Conference on Neural Information Processing Systems
🔥 [CVPR2024] Official implementation of "Self-correcting LLM-controlled Diffusion Models (SLD)
🔥🔥🔥A curated list of papers on recent diffusion-based high-resolution image and video synthesis works.
[NeurIPS 2024] ReNO: Enhancing One-step Text-to-Image Models through Reward-based Noise Optimization
[NeurIPS 2024] Official implementation of "BELM: Bidirectional Explicit Linear Multi-step Sampler for Exact Inversion in Diffusion Models".
Official Pytorch repo of CVPR'23 and NeurIPS'23 papers on understanding replication in diffusion models.
Scale-wise Distillation of Diffusion Models
🏞️ Official implementation of "Gen4Gen: Generative Data Pipeline for Generative Multi-Concept Composition"
[NeurIPS 2024] Official Implementation of Attention Interpolation of Text-to-Image Diffusion
[IJCAI 2025 (Oral)] Offical implementation of the paper "MagicTailor: Component-Controllable Personalization in Text-to-Image Diffusion Models".
A Large-scale Dataset for training and evaluating model's ability on Dense Text Image Generation
Official codebase for Margin-aware Preference Optimization for Aligning Diffusion Models without Reference (MaPO).
[IEEE TIP 2023] Txt2Img-MHN: Remote Sensing Image Generation from Text Using Modern Hopfield Networks
The largest multilingual image-text classification dataset. It contains fashion products.
[ECCV 2024 Oral] ConceptExpress: Harnessing Diffusion Models for Single-image Unsupervised Concept Extraction
[CVPR 2024] Tackling the Singularities at the Endpoints of Time Intervals in Diffusion Models
[ICLR 2024] Contextualized Diffusion Models for Text-Guided Image and Video Generation