starbeibei

starbeibei

Geek Repo

Github PK Tool:Github PK Tool

starbeibei's starred repositories

LAMM

[NeurIPS 2023 Datasets and Benchmarks Track] LAMM: Multi-Modal Large Language Models and Applications as AI Agents

Language:PythonStargazers:284Issues:0Issues:0

gv-benchmark

General Vision Benchmark, GV-B, a project from OpenGVLab

Language:PythonLicense:Apache-2.0Stargazers:187Issues:0Issues:0
Language:PythonStargazers:159Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:15Issues:0Issues:0
Language:PythonLicense:MITStargazers:69Issues:0Issues:0

HumanBench

This repo is official implementation of HumanBench (CVPR2023)

Language:PythonLicense:MITStargazers:216Issues:0Issues:0

VisionLLM

VisionLLM Series

Language:PythonLicense:Apache-2.0Stargazers:687Issues:0Issues:0
Stargazers:30Issues:0Issues:0

MUTR

[AAAI 2024] Referred by Multi-Modality: A Unified Temporal Transformers for Video Object Segmentation

Language:PythonLicense:MITStargazers:60Issues:0Issues:0

VideoMAEv2

[CVPR 2023] VideoMAE V2: Scaling Video Masked Autoencoders with Dual Masking

Language:PythonLicense:MITStargazers:444Issues:0Issues:0

M3I-Pretraining

[CVPR 2023] implementation of Towards All-in-one Pre-training via Maximizing Multi-modal Mutual Information.

Stargazers:87Issues:0Issues:0

CaFo

[CVPR 2023] Prompt, Generate, then Cache: Cascade of Foundation Models makes Strong Few-shot Learners

Language:PythonLicense:MITStargazers:332Issues:0Issues:0

GITM

Ghost in the Minecraft: Generally Capable Agents for Open-World Environments via Large Language Models with Text-based Knowledge and Memory

Stargazers:579Issues:0Issues:0
Stargazers:12Issues:0Issues:0

Instruct2Act

Instruct2Act: Mapping Multi-modality Instructions to Robotic Actions with Large Language Model

Language:PythonStargazers:299Issues:0Issues:0

UniHCP

Official PyTorch implementation of UniHCP

Language:PythonLicense:MITStargazers:140Issues:0Issues:0
Stargazers:17Issues:0Issues:0

Siamese-Image-Modeling

[CVPR 2023]Implementation of Siamese Image Modeling for Self-Supervised Vision Representation Learning

Language:PythonLicense:NOASSERTIONStargazers:32Issues:0Issues:0

Awesome-DragGAN

Awesome-DragGAN: A curated list of papers, tutorials, repositories related to DragGAN

Stargazers:82Issues:0Issues:0

DiffRate

[ICCV 23]An approach to enhance the efficiency of Vision Transformer (ViT) by concurrently employing token pruning and token merging techniques, while incorporating a differentiable compression rate.

Language:Jupyter NotebookStargazers:77Issues:0Issues:0

DragGAN

Unofficial Implementation of DragGAN - "Drag Your GAN: Interactive Point-based Manipulation on the Generative Image Manifold" (DragGAN 全功能实现,在线Demo,本地部署试用,代码、模型已全部开源,支持Windows, macOS, Linux)

Language:PythonStargazers:4994Issues:0Issues:0

UniFormerV2

[ICCV2023] UniFormerV2: Spatiotemporal Learning by Arming Image ViTs with Video UniFormer

Language:PythonLicense:Apache-2.0Stargazers:274Issues:0Issues:0

DDPS

Official Implementation of "Denoising Diffusion Semantic Segmentation with Mask Prior Modeling"

Language:PythonStargazers:63Issues:0Issues:0

Multitask-Model-Selector

Implementation of Foundation Model is Efficient Multimodal Multitask Model Selector

Language:PythonStargazers:28Issues:0Issues:0

Awesome-LLM4Tool

A curated list of the papers, repositories, tutorials, and anythings related to the large language models for tools

Stargazers:64Issues:0Issues:0

all-seeing

[ICLR 2024] This is the official implementation of the paper "The All-Seeing Project: Towards Panoptic Visual Recognition and Understanding of the Open World"

Language:PythonStargazers:411Issues:0Issues:0

InternGPT

InternGPT (iGPT) is an open source demo platform where you can easily showcase your AI models. Now it supports DragGAN, ChatGPT, ImageBind, multimodal chat like GPT-4, SAM, interactive image editing, etc. Try it at igpt.opengvlab.com (支持DragGAN、ChatGPT、ImageBind、SAM的在线Demo系统)

Language:PythonLicense:Apache-2.0Stargazers:3165Issues:0Issues:0

Ask-Anything

[CVPR2024 Highlight][VideoChatGPT] ChatGPT with video understanding! And many more supported LMs such as miniGPT4, StableLM, and MOSS.

Language:PythonLicense:MITStargazers:2851Issues:0Issues:0

OmniQuant

[ICLR2024 spotlight] OmniQuant is a simple and powerful quantization technique for LLMs.

Language:PythonLicense:MITStargazers:614Issues:0Issues:0