Zhixing Sun's repositories
Vitron
A Unified Pixel-level Vision LLM for Understanding, Generating, Segmenting, Editing
Monkey
【CVPR 2024 Highlight】Monkey (LMM): Image Resolution and Text Label Are Important Things for Large Multi-modal Models
IELT
Source code of the paper Fine-Grained Visual Classification via Internal Ensemble Learning Transformer
ovsam
[arXiv preprint] The official code of paper "Open-Vocabulary SAM".
Agent-Attention
Official repository of Agent Attention
FGVP
Official Codes for Fine-Grained Visual Prompting, NeurIPS 2023
2024-AAAI-HPT
Learning Hierarchical Prompt with Structured Linguistic Knowledge for Vision-Language Models (AAAI 2024)
LLaVA-Plus-Codebase
LLaVA-Plus: Large Language and Vision Assistants that Plug and Learn to Use Skills
MiniGPT-4
Open-sourced codes for MiniGPT-4 and MiniGPT-v2
FLatten-Transformer
Official repository of FLatten Transformer (ICCV2023)
RevisitingCIL
The code repository for "Revisiting Class-Incremental Learning with Pre-Trained Models: Generalizability and Adaptivity are All You Need" in PyTorch.
SHIP
Official code for ICCV 2023 paper, "Improving Zero-Shot Generalization for CLIP with Synthesized Prompts"
multimodal-prompt-learning
[CVPR 2023] Official repository of paper titled "MaPLe: Multi-modal Prompt Learning".
recognize-anything
Code for the Recognize Anything Model (RAM) and Tag2Text Model
AttriCLIP
CVPR2023: AttriCLIP: A Non-Incremental Learner for Incremental Knowledge Learning
code-samples
Holds code for our CVPR'23 tutorial: All Things ViTs: Understanding and Interpreting Attention in Vision.
sunhongbo.github.io
Github Pages template for academic personal websites, forked from mmistakes/minimal-mistakes
BiDistFSCIL
Official implementation of CVPR 2023 paper Few-Shot Class-Incremental Learning via Class-Aware Bilateral Distillation.
vit-pytorch
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
RPF
This is a repository contains the implementation of our SIGIR'23 full paper From Region to Patch: Attribute-Aware Foreground-Background Contrastive Learning for Fine-Grained Fashion Retrieval.
opencon
Code for TMLR 2023 paper "OpenCon: Open-world Contrastive Learning"
Gard
Code for Graph-based High-Order Relation Discovery for Fine-grained Recognition in CVPR 2021
APE
[ICCV 2023] Code for "Not All Features Matter: Enhancing Few-shot CLIP with Adaptive Prior Refinement"
CLIP_Surgery
CLIP Surgery for Better Explainability with Enhancement in Open-Vocabulary Tasks