bobzhang123's starred repositories
Grounded-Segment-Anything
Grounded SAM: Marrying Grounding DINO with Segment Anything & Stable Diffusion & Recognize Anything - Automatically Detect , Segment and Generate Anything
accelerate
🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support
Depth-Anything
[CVPR 2024] Depth Anything: Unleashing the Power of Large-Scale Unlabeled Data. Foundation Model for Monocular Depth Estimation
Awesome-Incremental-Learning
Awesome Incremental Learning
ICCV2023-Papers-with-Code
ICCV 2023 论文和开源项目合集
Awesome-Pruning
A curated list of neural network pruning resources.
occupancy_networks
This repository contains the code for the paper "Occupancy Networks - Learning 3D Reconstruction in Function Space"
Semi-supervised-learning
A Unified Semi-Supervised Learning Codebase (NeurIPS'22)
InternVideo
[ECCV2024] Video Foundation Models & Data for Multimodal Understanding
Awesome-Foundation-Models
A curated list of foundation models for vision and language tasks
SoraReview
The official GitHub page for the review paper "Sora: A Review on Background, Technology, Limitations, and Opportunities of Large Vision Models".
VideoMAEv2
[CVPR 2023] VideoMAE V2: Scaling Video Masked Autoencoders with Dual Masking
Awesome-Segment-Anything
A collection of project, papers, and source code for Meta AI's Segment Anything Model (SAM) and related studies.
Forge_VFM4AD
A comprehensive survey of forging vision foundation models for autonomous driving, including challenges, methodologies, and opportunities.
Awesome-Surrounding-Semantic-Occupancy-Prediction
Awesome papers about Multi-Camera Semantic Occupancy Prediction, such as TPVFormer, OccFormer, Occ3D, OpenOccupancy
awesome-video-self-supervised-learning
A curated list of awesome self-supervised learning methods in videos