Awj2021's repositories
awesome-labels-learning
The papers and projects with multi-label learning
diffusion-classifier
Diffusion Classifier leverages pretrained diffusion models to perform zero-shot classification without additional training
ProMix
PyTorch Code for ProMix: Combating Label Noise via Maximizing Clean Sample Utility
ControlNet
Let us control diffusion models!
da-fusion
Effective Data Augmentation With Diffusion Models
DiffMIC
[MICCAI 2023] DiffMIC: Dual-Guidance Diffusion Network for Medical Image Classification
diffusers
🤗 Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch and FLAX.
FROSTER
The official repository for ICLR2024 paper "FROSTER: Frozen CLIP is a Strong Teacher for Open-Vocabulary Action Recognition"
ICLR24
Official code for ICLR 2024 paper, "A Hard-to-Beat Baseline for Training-free CLIP-based Adaptation"
IP-Adapter
The image prompt adapter is designed to enable a pretrained text-to-image diffusion model to generate images with image prompt.
LRA-diffusion
This is the source code of LRA-diffusion for learning from noisy labels
MKT
Official implementation of "Open-Vocabulary Multi-Label Classification via Multi-Modal Knowledge Transfer".
open_clip
An open source implementation of CLIP.
PVT
Pyramid Transformer Networks for Our Own dataset.
quilt1m
[NeurIPS 2023 Oral] Quilt-1M: One Million Image-Text Pairs for Histopathology.
ResourceOfAI
Add some codes and commands which are common in research
SoTTA
This is the official PyTorch Implementation of "SoTTA: Robust Test-Time Adaptation on Noisy Data Streams (NeurIPS '23)" by Taesik Gong*, Yewon Kim*, Taeckyung Lee*, Sorn Chottananurak, and Sung-Ju Lee (* Equal contribution).
Swin-Transformer
This is an official implementation for "Swin Transformer: Hierarchical Vision Transformer using Shifted Windows".
taming-transformers
Taming Transformers for High-Resolution Image Synthesis
TestProjects
Mainly including some test files, like jupyter notebooks
tiny-transformers
[ECCV 2022] Implementation of the paper "Locality Guidance for Improving Vision Transformers on Tiny Datasets"
U-Mamba
U-Mamba: Enhancing Long-range Dependency for Biomedical Image Segmentation
unionnet
Replementation of unionnet "Deep Learning from Multiple Noisy Annotators as A Union"
Vim
Vision Mamba: Efficient Visual Representation Learning with Bidirectional State Space Model
VMamba
VMamba: Visual State Space Models
VPD
[ICCV 2023] VPD is a framework that leverages the high-level and low-level knowledge of a pre-trained text-to-image diffusion model to downstream visual perception tasks.