John Huang's repositories
Baichuan-7B
A large-scale 7B pretraining language model developed by BaiChuan-Inc.
Baichuan2
A series of large language models developed by Baichuan Intelligent Technology
open-interpreter
OpenAI's Code Interpreter in your terminal, running locally
chatglm.cpp
C++ implementation of ChatGLM-6B & ChatGLM2-6B & ChatGLM3 & more LLMs
ChatGLM3
ChatGLM3 series: Open Bilingual Chat LLMs | 开源双语对话语言模型
Chinese-LLaMA-Alpaca-2
中文LLaMA-2 & Alpaca-2大模型二期项目 + 16K超长上下文模型 (Chinese LLaMA-2 & Alpaca-2 LLMs, including 16K long context models)
codellama
Inference code for CodeLlama models
DyGLib
A Library for Dynamic Graph Learning (NeurIPS 2023)
FinGLM
FinGLM: 致力于构建一个开放的、公益的、持久的金融大模型项目,利用开源开放来促进「AI+金融」。
flash-attention
Fast and memory-efficient exact attention
gpt-fast
Simple and efficient pytorch-native transformer text generation in <1000 LOC of python.
Langchain-Chatchat
Langchain-Chatchat(原Langchain-ChatGLM)基于 Langchain 与 ChatGLM 等语言模型的本地知识库问答 | Langchain-Chatchat (formerly langchain-ChatGLM), local knowledge based LLM (like ChatGLM) QA app with langchain
lit-llama
Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
llama
Inference code for LLaMA models
LLaMA-Pro
Progressive LLaMA with Block Expansion.
llama.cpp
Port of Facebook's LLaMA model in C/C++
Llama2-Chinese
Llama中文社区,最好的中文Llama大模型,完全开源可商用
MemGPT
Teaching LLMs memory management for unbounded context 📚🦙
metaseq
Repo for external large-scale work
Qwen
The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud.
Qwen-VL
The official repo of Qwen-VL (通义千问-VL) chat & pretrained large vision language model proposed by Alibaba Cloud.
qwen.cpp
C++ implementation of Qwen-LM
TigerBot
TigerBot: A multi-language multi-task LLM
TinyLlama
The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.
trl
Train transformer language models with reinforcement learning.