Chaofan Tao's starred repositories
Megatron-LLaMA
Best practice for training LLaMA models in Megatron-LM
GPT-SoVITS
1 min voice data can also be used to train a good TTS model! (few shot voice cloning)
instruct-eval
This repository contains code to quantitatively evaluate instruction-tuned models such as Alpaca and Flan-T5 on held-out tasks.
Long-Context-Data-Engineering
Implementation of paper Data Engineering for Scaling Language Models to 128K Context
how-to-train-tokenizer
怎么训练一个LLM分词器
datablations
Scaling Data-Constrained Language Models
promptbench
A unified evaluation framework for large language models
alignment-handbook
Robust recipes to align language models with human and AI preferences
Megatron-LLM
distributed trainer for LLMs
Efficient-LLMs-Survey
[TMLR 2024] Efficient Large Language Models: A Survey
data-juicer
A one-stop data processing system to make data higher-quality, juicier, and more digestible for (multimodal) LLMs! 🍎 🍋 🌽 ➡️ ➡️🍸 🍹 🍷为大模型提供更高质量、更丰富、更易”消化“的数据!
OpenAgents
OpenAgents: An Open Platform for Language Agents in the Wild
visual_prompt_retrieval
[NeurIPS2023] Official implementation and model release of the paper "What Makes Good Examples for Visual In-Context Learning?"
Awesome-LLMs-Evaluation-Papers
The papers are organized according to our survey: Evaluating Large Language Models: A Comprehensive Survey.
LLaVA-RLHF
Aligning LMMs with Factually Augmented RLHF
generative-models
Generative Models by Stability AI
latent-consistency-model
Latent Consistency Models: Synthesizing High-Resolution Images with Few-Step Inference