hkxIron's starred repositories

sglang

SGLang is yet another fast serving framework for large language models and vision language models.

Language:PythonLicense:Apache-2.0Stargazers:4289Issues:0Issues:0

paper-reading

深度学习经典、新论文逐段精读

License:Apache-2.0Stargazers:25774Issues:0Issues:0

fairscale

PyTorch extensions for high performance and large scale training.

Language:PythonLicense:NOASSERTIONStargazers:3119Issues:0Issues:0

llama-models

Utilities intended for use with Llama models.

Language:PythonLicense:NOASSERTIONStargazers:3491Issues:0Issues:0
Language:PythonStargazers:151Issues:0Issues:0

zero_nlp

中文nlp解决方案(大模型、数据、模型、训练、推理)

Language:Jupyter NotebookLicense:MITStargazers:2792Issues:0Issues:0

accelerate

🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support

Language:PythonLicense:Apache-2.0Stargazers:7554Issues:0Issues:0

LLM-Pretrain-SFT

Scripts of LLM pre-training and fine-tuning (w/wo LoRA, DeepSpeed)

Language:PythonLicense:Apache-2.0Stargazers:61Issues:0Issues:0

LLM101n

LLM101n: Let's build a Storyteller

Stargazers:27162Issues:0Issues:0

SPACE

Official implementation of SPACE

Language:PythonLicense:Apache-2.0Stargazers:7Issues:0Issues:0

Spec-Bench

Spec-Bench: A Comprehensive Benchmark and Unified Evaluation Platform for Speculative Decoding (ACL 2024 Findings)

Language:PythonLicense:Apache-2.0Stargazers:142Issues:0Issues:0

COMET

A Neural Framework for MT Evaluation

Language:PythonLicense:Apache-2.0Stargazers:469Issues:0Issues:0

LMOps

General technology for enabling AI capabilities w/ LLMs and MLLMs

Language:PythonLicense:MITStargazers:3509Issues:0Issues:0

leedl-tutorial

《李宏毅深度学习教程》(李宏毅老师推荐👍,苹果书🍎),PDF下载地址:https://github.com/datawhalechina/leedl-tutorial/releases

Language:Jupyter NotebookLicense:NOASSERTIONStargazers:12552Issues:0Issues:0

MInference

To speed up Long-context LLMs' inference, approximate and dynamic sparse calculate the attention, which reduces inference latency by up to 10x for pre-filling on an A100 while maintaining accuracy.

Language:PythonLicense:MITStargazers:658Issues:0Issues:0

chat-dataset-baseline

人工精调的中文对话数据集和一段chatglm的微调代码

Language:Jupyter NotebookStargazers:1129Issues:0Issues:0

ChatGLM-Finetuning

基于ChatGLM-6B、ChatGLM2-6B、ChatGLM3-6B模型,进行下游具体任务微调,涉及Freeze、Lora、P-tuning、全参微调等

Language:PythonStargazers:2611Issues:0Issues:0

Cherry_LLM

[NAACL'24] Self-data filtering of LLM instruction-tuning data using a novel perplexity-based difficulty score, without using any other models

Language:PythonStargazers:258Issues:0Issues:0

LLMBook-zh.github.io

《大语言模型》作者:赵鑫,李军毅,周昆,唐天一,文继荣

Stargazers:2087Issues:0Issues:0

LLMTest_NeedleInAHaystack

Doing simple retrieval from LLM models at various context lengths to measure accuracy

Language:Jupyter NotebookLicense:NOASSERTIONStargazers:1400Issues:0Issues:0

mlmm-evaluation

Multilingual Large Language Models Evaluation Benchmark

Language:PythonLicense:Apache-2.0Stargazers:81Issues:0Issues:0
Language:PythonLicense:CC-BY-4.0Stargazers:276Issues:0Issues:0

llama3

The official Meta Llama 3 GitHub site

Language:PythonLicense:NOASSERTIONStargazers:25676Issues:0Issues:0

minbpe

Minimal, clean code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization.

Language:PythonLicense:MITStargazers:8943Issues:0Issues:0

Chinese-Mixtral

中文Mixtral混合专家大模型(Chinese Mixtral MoE LLMs)

Language:PythonLicense:Apache-2.0Stargazers:571Issues:0Issues:0

mistral-inference

Official inference library for Mistral models

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:9442Issues:0Issues:0

REST

REST: Retrieval-Based Speculative Decoding, NAACL 2024

Language:CLicense:Apache-2.0Stargazers:154Issues:0Issues:0

mamba

Mamba SSM architecture

Language:PythonLicense:Apache-2.0Stargazers:12232Issues:0Issues:0

Medusa

Medusa: Simple Framework for Accelerating LLM Generation with Multiple Decoding Heads

Language:Jupyter NotebookLicense:Apache-2.0Stargazers:2146Issues:0Issues:0

parallel-decoding

Repository of the paper "Accelerating Transformer Inference for Translation via Parallel Decoding"

Language:PythonLicense:Apache-2.0Stargazers:99Issues:0Issues:0