gaopengpjlab's starred repositories
segment-anything
The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
LLaMA-Adapter
[ICLR 2024] Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters
LLaMA2-Accessory
An Open-source Toolkit for LLM Development
Lumina-T2X
Lumina-T2X is a unified framework for Text to Any Modality Generation
Multi-Modality-Arena
Chatbot Arena meets multi-modality! Multi-Modality Arena allows you to benchmark vision-language models side-by-side while providing images as inputs. Supports MiniGPT-4, LLaMA-Adapter V2, LLaVA, BLIP-2, and many more!
Stable-Pix2Seq
A full-fledged version of Pix2Seq
PointCLIP_V2
[ICCV 2023] PointCLIP V2: Prompting CLIP and GPT for Powerful 3D Open-world Learning
Point-M2AE
[NeurIPS 2022] Multi-scale Masked Autoencoders for Hierarchical Point Cloud Pre-training
FeatAug-DETR
Official repository of paper: "FeatAug-DETR: Enriching One-to-Many Matching for DETRs with Feature Augmentation"
svl_adapter
SVL-Adapter: Self-Supervised Adapter for Vision-Language Pretrained Models
MonoDETR-MV
The multi-view version of MonoDETR on nuScenes dataset