Renrui Zhang (ZrrSkywalker)

ZrrSkywalker

Geek Repo

Company:CUHK MMLab

Location:Hong Kong

Home Page:https://zrrskywalker.github.io/

Github PK Tool:Github PK Tool

Renrui Zhang's starred repositories

segment-anything

The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:44612Issues:294Issues:640

Awesome-Multimodal-Large-Language-Models

:sparkles::sparkles:Latest Papers and Datasets on Multimodal Large Language Models, and Their Evaluation.

Transformers-Tutorials

This repository contains demos I made with the Transformers library by HuggingFace.

Language:Jupyter NotebookLicense:MITStargazers:8018Issues:124Issues:409

ImageBind

ImageBind One Embedding Space to Bind Them All

Language:PythonLicense:NOASSERTIONStargazers:7952Issues:99Issues:83

lit-llama

Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.

Language:PythonLicense:Apache-2.0Stargazers:5832Issues:68Issues:268

LLaMA-Adapter

[ICLR 2024] Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters

Language:PythonLicense:GPL-3.0Stargazers:5550Issues:78Issues:141

DragGAN

Unofficial Implementation of DragGAN - "Drag Your GAN: Interactive Point-based Manipulation on the Generative Image Manifold" (DragGAN 全功能实现,在线Demo,本地部署试用,代码、模型已全部开源,支持Windows, macOS, Linux)

Segment-Everything-Everywhere-All-At-Once

[NeurIPS 2023] Official implementation of the paper "Segment Everything Everywhere All at Once"

Language:PythonLicense:Apache-2.0Stargazers:4090Issues:55Issues:133

GPT-4-LLM

Instruction Tuning with GPT-4

Language:HTMLLicense:Apache-2.0Stargazers:4017Issues:46Issues:32

LLaMA2-Accessory

An Open-source Toolkit for LLM Development

Language:PythonLicense:NOASSERTIONStargazers:2559Issues:36Issues:128

Painter

Painter & SegGPT Series: Vision Foundation Models from BAAI

Language:PythonLicense:MITStargazers:2440Issues:36Issues:65

Personalize-SAM

Personalize Segment Anything Model (SAM) with 1 shot in 10 seconds

Language:PythonLicense:MITStargazers:1445Issues:27Issues:44

prolificdreamer

ProlificDreamer: High-Fidelity and Diverse Text-to-3D Generation with Variational Score Distillation (NeurIPS 2023 Spotlight)

Language:PythonLicense:Apache-2.0Stargazers:1396Issues:114Issues:21

LLaVA-Plus-Codebase

LLaVA-Plus: Large Language and Vision Assistants that Plug and Learn to Use Skills

Language:PythonLicense:Apache-2.0Stargazers:639Issues:10Issues:24

OneLLM

[CVPR 2024] OneLLM: One Framework to Align All Modalities with Language

Language:PythonLicense:NOASSERTIONStargazers:468Issues:11Issues:17

PointLLM

[arXiv 2023] PointLLM: Empowering Large Language Models to Understand Point Clouds

Point-Bind_Point-LLM

Align 3D Point Cloud with Multi-modalities for Large Language Models

Language:PythonLicense:MITStargazers:371Issues:14Issues:12

MonoDETR

[ICCV 2023] The first DETR model for monocular 3D object detection with depth-guided transformer

APE

[ICCV 2023] Code for "Not All Features Matter: Enhancing Few-shot CLIP with Adaptive Prior Refinement"

Language:Jupyter NotebookStargazers:125Issues:9Issues:12

lightning-GPT

Train and run GPTs with Lightning

Language:PythonLicense:Apache-2.0Stargazers:91Issues:14Issues:3

IAE

[ICCV 2023] "Implicit Autoencoder for Point-Cloud Self-Supervised Representation Learning"

MUTR

[AAAI 2024] Referred by Multi-Modality: A Unified Temporal Transformers for Video Object Segmentation

Language:PythonLicense:MITStargazers:56Issues:3Issues:3

ViewRefer3D

Official implementation of 'ViewRefer: Grasp the Multi-view Knowledge for 3D Visual Grounding with GPT and Prototype Guidance' (ICCV2023)

MV-JAR

[CVPR 2023] MV-JAR: Masked Voxel Jigsaw and Reconstruction for LiDAR-Based Self-Supervised Pre-Training

Point-PEFT

Point-PEFT: Parameter-Efficient Fine-Tuning for 3D Pre-trained Models(AAAI2024)

TFS3D

[CVPR2024] Less is more: Towards efficient few-shot 3D semantic segmentation via training-free networks

Language:PythonLicense:Apache-2.0Stargazers:6Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:4Issues:1Issues:1