Wizyoung's starred repositories

ChuanhuChatGPT

GUI for ChatGPT API and many LLMs. Supports agents, file-based QA, GPT finetuning and query with web search. All with a neat UI.

Language:PythonLicense:GPL-3.0Stargazers:14913Issues:85Issues:773

pelican

Static site generator that supports Markdown and reST syntax. Powered by Python.

Language:PythonLicense:AGPL-3.0Stargazers:12329Issues:337Issues:1656

Awesome-Chinese-LLM

整理开源的中文大语言模型,以规模较小、可私有化部署、训练成本较低的模型为主,包括底座模型,垂直领域微调及应用,数据集与教程等。

unsloth

Finetune Llama 3, Mistral, Phi & Gemma LLMs 2-5x faster with 80% less memory

Language:PythonLicense:Apache-2.0Stargazers:11076Issues:73Issues:446

GPT-4-LLM

Instruction Tuning with GPT-4

Language:HTMLLicense:Apache-2.0Stargazers:4036Issues:46Issues:33

AnyText

Official implementation code of the paper <AnyText: Multilingual Visual Text Generation And Editing>

Language:PythonLicense:Apache-2.0Stargazers:3894Issues:53Issues:97

mPLUG-DocOwl

mPLUG-DocOwl: Modularized Multimodal Large Language Model for Document Understanding

Language:PythonLicense:Apache-2.0Stargazers:1057Issues:27Issues:77

JetMoE

Reaching LLaMA2 Performance with 0.1M Dollars

Language:PythonLicense:Apache-2.0Stargazers:935Issues:8Issues:8

lmms-eval

Accelerating the development of large multimodal models (LMMs) with lmms-eval

Language:PythonLicense:NOASSERTIONStargazers:911Issues:3Issues:64

Bunny

A family of lightweight multimodal models.

Language:PythonLicense:Apache-2.0Stargazers:723Issues:20Issues:77

LLaVA-pp

🔥🔥 LLaVA++: Extending LLaVA with Phi-3 and LLaMA-3 (LLaVA LLaMA-3, LLaVA Phi-3)

Language:PythonLicense:MITStargazers:672Issues:6Issues:22

honeybee

Official implementation of project Honeybee (CVPR 2024)

Language:PythonLicense:NOASSERTIONStargazers:382Issues:15Issues:19

qwen-free-api

🚀 阿里通义千问2.5大模型逆向API白嫖测试【特长:六边形战士】,支持高速流式输出、无水印AI绘图、长文档解读、图像解析、多轮对话,零配置部署,多路token支持,自动清理会话痕迹。

Language:TypeScriptLicense:GPL-3.0Stargazers:364Issues:4Issues:28

Tabular-LLM

本项目旨在收集开源的表格智能任务数据集(比如表格问答、表格-文本生成等),将原始数据整理为指令微调格式的数据并微调LLM,进而增强LLM对于表格数据的理解,最终构建出专门面向表格智能任务的大型语言模型。

SEED-X

Multimodal Models in Real World

Language:Jupyter NotebookLicense:NOASSERTIONStargazers:275Issues:18Issues:11

MoAI

Official PyTorch implementation code for realizing the technical part of Mixture of All Intelligence (MoAI) to improve performance of numerous zero-shot vision language tasks. (Under Review)

Language:PythonLicense:MITStargazers:253Issues:9Issues:15

scaling_on_scales

When do we not need larger vision models?

Language:PythonLicense:MITStargazers:235Issues:3Issues:8

ChartVLM

Official Repository of ChartX & ChartVLM: A Versatile Benchmark and Foundation Model for Complicated Chart Reasoning

Language:PythonLicense:CC-BY-4.0Stargazers:183Issues:3Issues:11

FastV

Code for paper: An Image is Worth 1/2 Tokens After Layer 2: Plug-and-Play Inference Acceleration for Large Vision-Language Models

VILA

VILA - A multi-image visual language model with training, inference and evaluation recipe, deployable from cloud to edge (Jetson Orin and laptops)

Language:PythonLicense:Apache-2.0Stargazers:139Issues:9Issues:13

MathVerse

Does Your Multi-modal LLM Truly See the Diagrams in Visual Math Problems?

Language:PythonLicense:MITStargazers:109Issues:6Issues:4

G-LLaVA

Official github repo of G-LLaVA

CuMo

CuMo: Scaling Multimodal LLM with Co-Upcycled Mixture-of-Experts

Language:PythonLicense:Apache-2.0Stargazers:92Issues:1Issues:9
Language:PythonLicense:Apache-2.0Stargazers:14Issues:2Issues:1

mamba-gpt-3b

It is almost the best 3B model in the current open source industry, surpassing Dolly v2-3b, open lama-3b, and even outperforming the EleutherAI/pythia-12b model in terms of performance. Can refer to open_llm_leaderboard

Language:PythonLicense:MITStargazers:10Issues:1Issues:0