PL's repositories
ALIA
Augmenting with Language-guided Image Augmentation
Ask-Anything
[VideoChatGPT] ChatGPT with video understanding! And many more supported LMs such as miniGPT4, StableLM, and MOSS.
Awesome_Prompting_Papers_in_Computer_Vision
A curated list of prompt-based paper in computer vision and vision-language learning.
CLCAE
Delving StyleGAN Inversion for Image Editing: A Foundation Latent Space Viewpoint
Clip2Protect
[CVPR 2023] Official repository of paper titled "CLIP2Protect: Protecting Facial Privacy using Text-Guided Makeup via Adversarial Latent Search".
CUDA_LTR
Official Implementation of Curriculum of Data Augmentation for Long-tailed Recognition (CUDA) (ICLR'23 Spotlight)
DiLM
Implementaiton of "DiLM: Distilling Dataset into Language Model for Text-level Dataset Distillation" (accepted by NAACL2024 Findings)".
EAT_code
Official code for ICCV 2023 paper: "Efficient Emotional Adaptation for Audio-Driven Talking-Head Generation".
Efficient-Dataset-Condensation
Official PyTorch implementation of "Dataset Condensation via Efficient Synthetic-Data Parameterization" (ICML'22)
FATE-LLM
Federated Learning for LLMs.
FreD
Official PyTorch implementation for Frequency Domain-based Dataset Distillation [NeurIPS 2023]
generative_agents
Generative Agents: Interactive Simulacra of Human Behavior
ILM-VP
[CVPR23] "Understanding and Improving Visual Prompting: A Label-Mapping Perspective" by Aochuan Chen, Yuguang Yao, Pin-Yu Chen, Yihua Zhang, and Sijia Liu
LLaMA-Adapter
Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters
LLaVA
Large Language-and-Vision Assistant built towards multimodal GPT-4 level capabilities.
Otter
🦦 Otter, a multi-modal model based on OpenFlamingo (open-sourced version of DeepMind's Flamingo), trained on MIMIC-IT and showcasing improved instruction-following and in-context learning ability.
pingliu264.github.io
Github Pages template for academic personal websites, forked from mmistakes/minimal-mistakes
PoDD
Official PyTorch Implementation for the "Distilling Datasets Into Less Than One Image" paper.
Point-In-Context
Implementation of the paper: Explore In-Context Learning for 3D Point Cloud Understanding
SRe2L
Large-scale Dataset Distillation/Condensation, 50 IPC (Images Per Class) achieves highest 60.8% on original ImageNet-1K val set.
stable-diffusion
A latent text-to-image diffusion model
stanford_alpaca
Code and documentation to train Stanford's Alpaca models, and generate the data.
Video-LLaMA
Video-LLaMA: An Instruction-tuned Audio-Visual Language Model for Video Understanding
video_distillation
Official implementation of Dancing with Still Images: Video Distillation via Static-Dynamic Disentanglement.
WatermarkDM
Code of the paper: A Recipe for Watermarking Diffusion Models