Mark Peng's starred repositories
Depth-Anything
[CVPR 2024] Depth Anything: Unleashing the Power of Large-Scale Unlabeled Data. Foundation Model for Monocular Depth Estimation
Inpaint-Anything
Inpaint anything using Segment Anything and inpainting models.
LLM-Agent-Paper-List
The paper list of the 86-page paper "The Rise and Potential of Large Language Model Based Agents: A Survey" by Zhiheng Xi et al.
GroundingDINO
Official implementation of the paper "Grounding DINO: Marrying DINO with Grounded Pre-Training for Open-Set Object Detection"
alignment-handbook
Robust recipes to align language models with human and AI preferences
Deformable-DETR
Deformable DETR: Deformable Transformers for End-to-End Object Detection.
EfficientSAM
EfficientSAM: Leveraged Masked Image Pretraining for Efficient Segment Anything
unlimiformer
Public repo for the NeurIPS 2023 paper "Unlimiformer: Long-Range Transformers with Unlimited Length Input"
Awesome-Open-Vocabulary
(TPAMI 2024) A Survey on Open Vocabulary Learning
Segment-Any-Anomaly
Official implementation of "Segment Any Anomaly without Training via Hybrid Prompt Regularization (SAA+)".
linear-attention-transformer
Transformer based on a variant of attention that is linear complexity in respect to sequence length
Open-GroundingDino
This is the third party implementation of the paper Grounding DINO: Marrying DINO with Grounded Pre-Training for Open-Set Object Detection.
EEG-ATCNet
Attention temporal convolutional network for EEG-based motor imagery classification
LearnablePromptSAM
Try to use the SAM-ViT as the backbone to create the learnable prompt for semantic segmentation