Mingdeng's starred repositories
segment-anything
The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
AnimateAnyone
Animate Anyone: Consistent and Controllable Image-to-Video Synthesis for Character Animation
PhotoMaker
PhotoMaker
VideoCrafter
VideoCrafter2: Overcoming Data Limitations for High-Quality Video Diffusion Models
T2I-Adapter
T2I-Adapter
Caption-Anything
Caption-Anything is a versatile tool combining image segmentation, visual captioning, and ChatGPT, generating tailored captions with diverse controls for user preferences. https://huggingface.co/spaces/TencentARC/Caption-Anything https://huggingface.co/spaces/VIPLab/Caption-Anything
MotionCtrl
MotionCtrl: A Unified and Flexible Motion Controller for Video Generation
StableDiffusionReconstruction
Takagi and Nishimoto, CVPR 2023
ScaleCrafter
[ICLR 2024 Spotlight] Official implementation of ScaleCrafter for higher-resolution visual generation at inference time.
Mix-of-Show
NeurIPS 2023, Mix-of-Show: Decentralized Low-Rank Adaptation for Multi-Concept Customization of Diffusion Models
InterpAny-Clearer
Clearer anytime frame interpolation & Manipulated interpolation of anything
360SR-Challenge
NTIRE 2023 Challenge on 360° Omnidirectional Image and Video Super-Resolution
Assessor360
[NeurIPS 2023] Assessor360: Multi-sequence Network for Blind Omnidirectional Image Quality Assessment
T2I-Adapter-w-MasaCtrl
MasaCtrl with T2I-Adapter for controllable consistent image synthesis and editing