ScottWang's starred repositories
InternLM-Math
State-of-the-art bilingual open-sourced Math reasoning LLMs.
lm-human-preferences
Code for the paper Fine-Tuning Language Models from Human Preferences
awesome-instruction-datasets
A collection of awesome-prompt-datasets, awesome-instruction-dataset, to train ChatLLM such as chatgpt 收录各种各样的指令数据集, 用于训练 ChatLLM 模型。
instruction-datasets
All available datasets for Instruction Tuning of Large Language Models
AgentBench
A Comprehensive Benchmark to Evaluate LLMs as Agents (ICLR'24)
LLaMA-Factory
Efficiently Fine-Tune 100+ LLMs in WebUI (ACL 2024)
BIG-Bench-Hard
Challenging BIG-Bench Tasks and Whether Chain-of-Thought Can Solve Them
direct-preference-optimization
Reference implementation for DPO (Direct Preference Optimization)
Awesome-Chinese-LLM
整理开源的中文大语言模型,以规模较小、可私有化部署、训练成本较低的模型为主,包括底座模型,垂直领域微调及应用,数据集与教程等。
chain-of-thought-hub
Benchmarking large language models' complex reasoning ability with chain-of-thought prompting
Safety-Prompts
Chinese safety prompts for evaluating and improving the safety of LLMs. 中文安全prompts,用于评估和提升大模型的安全性。
alpaca-lora
Instruct-tune LLaMA on consumer hardware
DeepSpeedExamples
Example models using DeepSpeed
ChatGLM-6B
ChatGLM-6B: An Open Bilingual Dialogue Language Model | 开源双语对话语言模型
stanford_alpaca
Code and documentation to train Stanford's Alpaca models, and generate the data.