suolyer's starred repositories
FlashAttention20Triton
Triton implementation of Flash Attention2.0
Awesome-LLM-Long-Context-Modeling
📰 Must-read papers and blogs on LLM based Long Context Modeling 🔥
long-llms-learning
A repository sharing the literatures about long-context large language models, including the methodologies and the evaluation benchmarks
triton_flashv2_alibi
working repo for Triton based Flash2 supporting alibi pos embeddings
synthesizer
A multi-purpose LLM framework for RAG and data creation.
textbook_quality
Generate textbook-quality synthetic LLM pretraining data
data-juicer
A one-stop data processing system to make data higher-quality, juicier, and more digestible for LLMs! 🍎 🍋 🌽 ➡️ ➡️🍸 🍹 🍷为大语言模型提供更高质量、更丰富、更易”消化“的数据!
text-dedup
Python package for memory-friendly text de-duplication
LLMDataHub
A quick guide (especially) for trending instruction finetuning datasets
Arxiv-NLP-Reporter
每日自动获取Arxiv上NLP相关最新论文【Arxiv Natural Language Processing Paper Automatic Crawl Daily】
sft_datasets
开源SFT数据集整理,随时补充
Cornucopia-LLaMA-Fin-Chinese
聚宝盆(Cornucopia): 中文金融系列开源可商用大模型,并提供一套高效轻量化的垂直领域LLM训练框架(Pretraining、SFT、RLHF、Quantize等)
WanJuan1.0
万卷1.0多模态语料
Finetune_LLAMA
简单易懂的LLaMA微调指南。
literature-books
书籍txt
High-quality-Chinese-Q-A-dataset
最大开源中文问答数据集 ,助力中文LLM.The largest open-source Chinese Q&A dataset, supporting Chinese LLM
Awesome-Chinese-LLM
整理开源的中文大语言模型,以规模较小、可私有化部署、训练成本较低的模型为主,包括底座模型,垂直领域微调及应用,数据集与教程等。
Open-Llama
The complete training code of the open-source high-performance Llama model, including the full process from pre-training to RLHF.
flash-attention
Fast and memory-efficient exact attention