QinLuo's starred repositories
streaming-llm
[ICLR 2024] Efficient Streaming Language Models with Attention Sinks
llm-foundry
LLM training code for Databricks foundation models
h2o-llmstudio
H2O LLM Studio - a framework and no-code GUI for fine-tuning LLMs. Documentation: https://h2oai.github.io/h2o-llmstudio/
state-of-open-source-ai
:closed_book: Clarity in the current fast-paced mess of Open Source innovation
Skywork
Skywork series models are pre-trained on 3.2TB of high-quality multilingual (mainly Chinese and English) and code data. We have open-sourced the model, training data, evaluation data, evaluation methods, etc. 天工系列模型在3.2TB高质量多语言和代码数据上进行预训练。我们开源了模型参数,训练数据,评估数据,评估方法。
local-persist
Create named local volumes that persist in the location(s) you want
rl_a3c_pytorch
A3C LSTM Atari with Pytorch plus A3G design
textbook_quality
Generate textbook-quality synthetic LLM pretraining data
adept-inference
Inference code for Persimmon-8B
libgen_to_txt
Convert all of libgen to high quality markdown
AttentionIsOFFByOne
Implementation of "Attention Is Off By One" by Evan Miller
multipack_sampler
Multipack distributed sampler for fast padding-free training of LLMs
AutoNetGen
让 AI 设计 AI,让大模型帮助小模型进化,用魔法创造魔法! Empower Artificial Intelligence to sculpt its own kind, where colossal models gracefully usher the petite ones into evolution, weaving magic to conjure further enchantment!
colorbindiff
A visual and colorized diff for binary files.
tqdm-loggable
Logging friendly progress messages for TQDM progress bars