Hay Kim (TongHengcheng)

TongHengcheng

Geek Repo

Company:Aire

Github PK Tool:Github PK Tool

Hay Kim's repositories

flux

Official inference repo for FLUX.1 models

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

MimicMotion

High-Quality Human Motion Video Generation with Confidence-aware Pose Guidance

Language:PythonLicense:NOASSERTIONStargazers:0Issues:0Issues:0

Awesome-Video-Diffusion

A curated list of recent diffusion models for video generation, editing, restoration, understanding, etc.

Stargazers:0Issues:0Issues:0

CogVideo

Text-to-video generation: CogVideoX (2024) and CogVideo (ICLR 2023)

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

AnyV2V

A Plug-and-Play Framework For Any Video-to-Video Editing Tasks

Language:Jupyter NotebookLicense:MITStargazers:0Issues:0Issues:0

champ

Champ: Controllable and Consistent Human Image Animation with 3D Parametric Guidance

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

ControlNeXt

Controllable video and image Generation, SVD, Animate Anyone, ControlNet, ControlNeXt, LoRA

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

GOT-OCR2.0-

Official code implementation of General OCR Theory: Towards OCR-2.0 via a Unified End-to-end Model

Language:PythonStargazers:0Issues:0Issues:0

Cinemo

Cinemo: Consistent and Controllable Image Animation with Motion Diffusion Models

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

FollowYourEmoji

[Siggraph Asia 2024] Follow-Your-Emoji: This repo is the official implementation of "Follow-Your-Emoji: Fine-Controllable and Expressive Freestyle Portrait Animation"

Stargazers:0Issues:0Issues:0

MOFA-Video

[ECCV 2024] MOFA-Video: Controllable Image Animation via Generative Motion Field Adaptions in Frozen Image-to-Video Diffusion Model.

Language:PythonLicense:NOASSERTIONStargazers:0Issues:0Issues:0

Show-o

Repository for Show-o, One Single Transformer to Unify Multimodal Understanding and Generation.

License:MITStargazers:0Issues:0Issues:0

Monkey

【CVPR 2024】Monkey (LMM): Image Resolution and Text Label Are Important Things for Large Multi-modal Models

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

manga-image-translator

Translate manga/image 一键翻译各类图片内文字 https://cotrans.touhou.ai/

License:GPL-3.0Stargazers:0Issues:0Issues:0

EchoMimic

Lifelike Audio-Driven Portrait Animations through Editable Landmark Conditioning

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

VEnhancer

Official codes of VEnhancer: Generative Space-Time Enhancement for Video Generation

Stargazers:0Issues:0Issues:0

Lumina-mGPT

Official Implementation of "Lumina-mGPT: Illuminate Flexible Photorealistic Text-to-Image Generation with Multimodal Generative Pretraining"

Stargazers:0Issues:0Issues:0

UniAnimate

Code for Paper "UniAnimate: Taming Unified Video Diffusion Models for Consistent Human Image Animation".

Language:PythonStargazers:0Issues:0Issues:0

mPLUG-Owl

mPLUG-Owl: The Powerful Multi-modal Large Language Model Family

License:MITStargazers:0Issues:0Issues:0

UniPortrait

UniPortrait: A Unified Framework for Identity-Preserving Single- and Multi-Human Image Personalizations

License:Apache-2.0Stargazers:0Issues:0Issues:0

Deep-Live-Cam

real time face swap and one-click video deepfake with only a single image (uncensored)

License:AGPL-3.0Stargazers:0Issues:0Issues:0

LivePortrait

Make one portrait alive!

Language:PythonLicense:NOASSERTIONStargazers:0Issues:0Issues:0

ComfyUI-RefUNet

A set of nodes to use Reference UNets

License:GPL-3.0Stargazers:0Issues:0Issues:0

SimpleTuner

A general fine-tuning kit geared toward Stable Diffusion 2.1, Stable Diffusion 3, DeepFloyd, and SDXL.

Language:PythonLicense:AGPL-3.0Stargazers:0Issues:0Issues:0

MiniCPM-V

MiniCPM-Llama3-V 2.5: A GPT-4V Level MLLM on Your Phone

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

OpenDiT

OpenDiT: An Easy, Fast and Memory-Efficient System for DiT Training and Inference

License:Apache-2.0Stargazers:0Issues:0Issues:0

segment-anything-2

The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.

License:Apache-2.0Stargazers:0Issues:0Issues:0

MagicClothing

Official implementation of Magic Clothing: Controllable Garment-Driven Image Synthesis

License:NOASSERTIONStargazers:0Issues:0Issues:0

HunyuanDiT

Hunyuan-DiT : A Powerful Multi-Resolution Diffusion Transformer with Fine-Grained Chinese Understanding

License:NOASSERTIONStargazers:0Issues:0Issues:0

MotionBooth

The official implement of research paper "MotionBooth: Motion-Aware Customized Text-to-Video Generation"

Language:PythonStargazers:0Issues:0Issues:0