Wei Zhou's starred repositories
open-interpreter
A natural language interface for computers
awesome-english-ebooks
经济学人(含音频)、纽约客、卫报、连线、大西洋月刊等英语杂志免费下载,支持epub、mobi、pdf格式, 每周更新
clip-as-service
🏄 Scalable embedding, reasoning, ranking for images and sentences with CLIP
open_llama
OpenLLaMA, a permissively licensed open source reproduction of Meta AI’s LLaMA 7B trained on the RedPajama dataset
Video-LLaMA
[EMNLP 2023 Demo] Video-LLaMA: An Instruction-tuned Audio-Visual Language Model for Video Understanding
Semantic-SAM
Official implementation of the paper "Semantic-SAM: Segment and Recognize Anything at Any Granularity"
autodistill
Images to inference with no labeling (use foundation models to train supervised models).
efficientvit
EfficientViT is a new family of vision models for efficient high-resolution vision.
awesome_lists
Awesome Lists for Tenure-Track Assistant Professors and PhD students. (助理教授/博士生生存指南)
lm-format-enforcer
Enforce the output format (JSON Schema, Regex etc) of a language model
Image2Paragraph
[A toolbox for fun.] Transform Image into Unique Paragraph with ChatGPT, BLIP2, OFA, GRIT, Segment Anything, ControlNet.
LLM-in-Vision
Recent LLM-based CV and related works. Welcome to comment/contribute!
COCO-WholeBody
ECCV2020 paper "Whole-Body Human Pose Estimation in the Wild"
Chinese-LLaVA
支持中英文双语视觉-文本对话的开源可商用多模态模型。
Awesome-Segment-Anything
A collection of project, papers, and source code for Meta AI's Segment Anything Model (SAM) and related studies.
LRV-Instruction
[ICLR'24] Mitigating Hallucination in Large Multi-Modal Models via Robust Instruction Tuning
Visual-Instruction-Tuning
SVIT: Scaling up Visual Instruction Tuning