jiangziben

jiangziben

Geek Repo

Github PK Tool:Github PK Tool

jiangziben's starred repositories

data-juicer

A one-stop data processing system to make data higher-quality, juicier, and more digestible for (multimodal) LLMs! 🍎 🍋 🌽 ➡️ ➡️🍸 🍹 🍷为大模型提供更高质量、更丰富、更易”消化“的数据!

Language:PythonLicense:Apache-2.0Stargazers:2260Issues:0Issues:0

fine-tune-train_segment_anything_2_in_60_lines_of_code

The repository provides code for training/fine tune the Meta Segment Anything Model 2 (SAM 2)

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:37Issues:0Issues:0

segment-anything-2

The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:9564Issues:0Issues:0

recognize-anything

Open-source and strong foundation image recognition models.

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:2686Issues:0Issues:0

cleanlab

The standard data-centric AI package for data quality and machine learning with messy, real-world data and labels.

Language:PythonLicense:AGPL-3.0Stargazers:9267Issues:0Issues:0

Bunny

A family of lightweight multimodal models.

Language:PythonLicense:Apache-2.0Stargazers:853Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:1902Issues:0Issues:0

Grounded-SAM-2

Grounded SAM 2: Ground and Track Anything in Videos with Grounding DINO and SAM 2

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:421Issues:0Issues:0

Deformable-DETR

Deformable DETR: Deformable Transformers for End-to-End Object Detection.

Language:PythonLicense:Apache-2.0Stargazers:3096Issues:0Issues:0

TAPTR

[ECCV 2024] Official implementation of the paper "TAPTR: Tracking Any Point with Transformers as Detection"

Language:PythonLicense:NOASSERTIONStargazers:169Issues:0Issues:0

Semantic-SAM

[ECCV 2024] Official implementation of the paper "Semantic-SAM: Segment and Recognize Anything at Any Granularity"

Language:PythonStargazers:2206Issues:0Issues:0

co-tracker

CoTracker is a model for tracking any point (pixel) on a video.

Language:Jupyter NotebookLicense:NOASSERTIONStargazers:2628Issues:0Issues:0

LLaVA

[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.

Language:PythonLicense:Apache-2.0Stargazers:18754Issues:0Issues:0

openpose

OpenPose: Real-time multi-person keypoint detection library for body, face, hands, and foot estimation

Language:C++License:NOASSERTIONStargazers:30632Issues:0Issues:0

anylabeling

Effortless AI-assisted data labeling with AI support from YOLO, Segment Anything (SAM+SAM2), MobileSAM!!

Language:PythonLicense:GPL-3.0Stargazers:2140Issues:0Issues:0

SAM-6D

[CVPR2024] Code for "SAM-6D: Segment Anything Model Meets Zero-Shot 6D Object Pose Estimation".

Language:PythonStargazers:298Issues:0Issues:0

yolov5

YOLOv5 🚀 in PyTorch > ONNX > CoreML > TFLite

Language:PythonLicense:AGPL-3.0Stargazers:49201Issues:0Issues:0

Mask2Former

Code release for "Masked-attention Mask Transformer for Universal Image Segmentation"

Language:PythonLicense:MITStargazers:2413Issues:0Issues:0

OpenScene

3D Occupancy Prediction Benchmark in Autonomous Driving

Language:PythonLicense:Apache-2.0Stargazers:284Issues:0Issues:0

ComfyUI

The most powerful and modular diffusion model GUI, api and backend with a graph/nodes interface.

Language:PythonLicense:GPL-3.0Stargazers:46985Issues:0Issues:0

bevfusion

[ICRA'23] BEVFusion: Multi-Task Multi-Sensor Fusion with Unified Bird's-Eye View Representation

Language:PythonLicense:Apache-2.0Stargazers:2191Issues:0Issues:0

fastdup

fastdup is a powerful free tool designed to rapidly extract valuable insights from your image & video datasets. Assisting you to increase your dataset images & labels quality and reduce your data operations costs at an unparalleled scale.

Language:PythonLicense:NOASSERTIONStargazers:1555Issues:0Issues:0

LLM101n

LLM101n: Let's build a Storyteller

Stargazers:27011Issues:0Issues:0

Eureka

Official Repository for "Eureka: Human-Level Reward Design via Coding Large Language Models" (ICLR 2024)

Language:Jupyter NotebookLicense:MITStargazers:2759Issues:0Issues:0

objectdetection_script

一些关于目标检测的脚本的改进思路代码,详细请看readme.md

Language:PythonStargazers:4942Issues:0Issues:0

pytorch-grad-cam

Advanced AI Explainability for computer vision. Support for CNNs, Vision Transformers, Classification, Object detection, Segmentation, Image similarity and more.

Language:PythonLicense:MITStargazers:10042Issues:0Issues:0

CameraLaserCalibrate

Camera Laser Calibrate python version

Language:PythonLicense:MITStargazers:31Issues:0Issues:0

cvat

Annotate better with CVAT, the industry-leading data engine for machine learning. Used and trusted by teams at any scale, for data of any scale.

Language:TypeScriptLicense:MITStargazers:12047Issues:0Issues:0

universal_manipulation_interface

Universal Manipulation Interface: In-The-Wild Robot Teaching Without In-The-Wild Robots

Language:PythonLicense:MITStargazers:562Issues:0Issues:0

umi-on-legs

UMI on Legs: Making Manipulation Policies Mobile with Manipulation-Centric Whole-body Controllers

Language:PythonLicense:MITStargazers:154Issues:0Issues:0