Fangkai Jiao's starred repositories
LLaMA-Factory
Unify Efficient Fine-Tuning of 100+ LLMs
DeepSeek-Coder
DeepSeek Coder: Let the Code Write Itself
DeepSpeed-MII
MII makes low-latency and high-throughput inference possible, powered by DeepSpeed.
TransformerEngine
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
LLMAgentPapers
Must-read Papers on LLM Agents.
DeepSeek-MoE
DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models
alpaca_farm
A simulation framework for RLHF and alternatives. Develop your RLHF method without collecting human data.
flash-linear-attention
Efficient implementations of state-of-the-art linear attention models in Pytorch and Triton
Awesome-LLMs-Evaluation-Papers
The papers are organized according to our survey: Evaluating Large Language Models: A Comprehensive Survey.
Awesome-Reasoning-Foundation-Models
✨✨Latest Papers and Benchmarks in Reasoning with Foundation Models
Q-Instruct
②[CVPR 2024] Low-level visual instruction tuning, with a 200K dataset and a model zoo for fine-tuned checkpoints.
OpenSource-LLMs-better-than-OpenAI
Listing all reported open-source LLMs achieving a higher score than proprietary, paying OpenAI models (ChatGPT, GPT-4).
gpt_paper_assistant
GPT4 based personalized ArXiv paper assistant bot