Manxi Lin's starred repositories
segment-anything-2
The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
Mask2Former
Code release for "Masked-attention Mask Transformer for Universal Image Segmentation"
InternImage
[CVPR 2023 Highlight] InternImage: Exploring Large-Scale Vision Foundation Models with Deformable Convolutions
Semantic-SAM
[ECCV 2024] Official implementation of the paper "Semantic-SAM: Segment and Recognize Anything at Any Granularity"
ViT-Adapter
[ICLR 2023 Spotlight] Vision Transformer Adapter for Dense Predictions
awesome_LLMs_interview_notes
LLMs interview notes and answers:该仓库主要记录大模型(LLMs)算法工程师相关的面试题和参考答案
Grounding-DINO-1.5-API
API for Grounding DINO 1.5: IDEA Research's Most Capable Open-World Object Detection Model Series
unmasked_teacher
[ICCV2023 Oral] Unmasked Teacher: Towards Training-Efficient Video Foundation Models
SurgicalDINO
[IPCAI'2024 (IJCARS special issue)] Surgical-DINO: Adapter Learning of Foundation Models for Depth Estimation in Endoscopic Surgery
Mask2Former_DINOv2
将Mask2Former的backbone替换成DINOv2训练好的ViT模型
CLIP-spurious-finetune
Mitigating Spurious Correlations in Multi-modal Models during Fine-tuning (ICML 2023)
OutlierDetectionChallenge2024
Outlier detection challenge 2024 - a DTU Compute summer school challenge
act-plus-plus
Imitation learning algorithms with Co-training for Mobile ALOHA: ACT, Diffusion Policy, VINN