Shwan's repositories
FindTheChatGPTer
汇总那些ChatGPT的平替们
alpaca-lora
Instruct-tune LLaMA on consumer hardware
Auto-GPT
An experimental open-source attempt to make GPT-4 fully autonomous.
BELLE
BELLE: Be Everyone's Large Language model Engine(开源中文对话大模型)
ChatGLM-6B
ChatGLM-6B:开源双语对话语言模型 | An Open Bilingual Dialogue Language Model
Chinese-alpaca-lora
骆驼:A Chinese finetuned instruction LLaMA. Developed by 陈启源 @ 华中师范大学 & 李鲁鲁 @ 商汤科技 & 冷子昂 @ 商汤科技
Chinese-ChatLLaMA
中文LLaMA基础模型;中文ChatLLaMA对话模型;NLP预训练/指令微调数据集
Chinese-LangChain
中文langchain项目|小必应,Q.Talk,强聊,QiangTalk
Chinese-Vicuna
Chinese-Vicuna: A Chinese Instruction-following LLaMA-based Model —— 一个中文低资源的llama+lora方案,结构参考alpaca
ColossalAI
Making large AI models cheaper, faster and more accessible
DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
dolly
Databricks’ Dolly, a large language model trained on the Databricks Machine Learning Platform
EasyLM
Large language models (LLMs) made easy, EasyLM is a one stop solution for pre-training, finetuning, evaluating and serving LLMs in JAX/Flax.
FastChat
The release repo for "Vicuna: An Open Chatbot Impressing GPT-4"
GLM-130B
GLM-130B: An Open Bilingual Pre-Trained Model (ICLR 2023)
GPTQ-for-LLaMa
4 bits quantization of LLaMA using GPTQ
langtorch
Building composable LLM applications with Java / JVM.
LMFlow
An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Model for All.
Megatron-LM
Ongoing research training transformer models at scale
nebullvm
Plug and play modules to optimize the performances of your AI systems 🚀
Open-Llama
The complete training code of the open-source high-performance Llama model, including the full process from pre-training to RLHF.
open_flamingo
An open-source framework for training large multimodal models.
Plan4MC
Reinforcement learning and planning for Minecraft.
StableLM
StableLM: Stability AI Language Models
stanford_alpaca
Code and documentation to train Stanford's Alpaca models, and generate the data.
transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
trlx
A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)
unit-minions
《AI 研发提效研究:自己动手训练 LoRA》,包含 Llama (Alpaca LoRA)模型、ChatGLM (ChatGLM Tuning)相关 Lora 的训练。训练内容:用户故事生成、测试代码生成、代码辅助生成、文本转 SQL、文本生成代码……