Gyeongchan-Yun's starred repositories

Awesome_LLM_System-PaperList

Since the emergence of chatGPT in 2022, the acceleration of Large Language Model has become increasingly important. Here is a list of papers on accelerating LLMs, currently focusing mainly on inference acceleration, and related works will be gradually added in the future. Welcome contributions!

Stargazers:117Issues:0Issues:0

Megatron-Kwai

Ongoing research training transformer models at scale

Language:PythonLicense:NOASSERTIONStargazers:14Issues:0Issues:0

AMP

(NeurIPS 2022) Automatically finding good model-parallel strategies, especially for complex models and clusters.

Language:PythonStargazers:32Issues:0Issues:0

alpa

Training and serving large-scale neural networks with auto parallelization.

Language:PythonLicense:Apache-2.0Stargazers:3013Issues:0Issues:0

Megatron-LM

Artifact for DynaPipe: Optimizing Multi-task Training through Dynamic Pipelines

Language:PythonLicense:NOASSERTIONStargazers:1Issues:0Issues:0

python-patterns

A collection of design patterns/idioms in Python

Language:PythonStargazers:39910Issues:0Issues:0

awesome-AI-system

paper and its code for AI System

Stargazers:187Issues:0Issues:0

zero-bubble-pipeline-parallelism

Zero Bubble Pipeline Parallelism

Language:PythonLicense:NOASSERTIONStargazers:228Issues:0Issues:0

optimizing-multitask-training-through-dynamic-pipelines

Official repository for the paper DynaPipe: Optimizing Multi-task Training through Dynamic Pipelines

Language:PythonLicense:Apache-2.0Stargazers:11Issues:0Issues:0

Optimus-CC

[ASPLOS'23] Optimus-CC: Efficient Large NLP Model Training with 3D Parallelism Aware Communication Compression

License:MITStargazers:3Issues:0Issues:0

Hetu-Galvatron

Galvatron is an automatic distributed training system designed for Transformer models, including Large Language Models (LLMs). If you have any interests, please visit/star/fork https://github.com/PKU-DAIR/Hetu-Galvatron

Language:PythonStargazers:9Issues:0Issues:0
Language:PythonStargazers:5Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:242Issues:0Issues:0
Language:Jupyter NotebookLicense:CC-BY-4.0Stargazers:111Issues:0Issues:0

HexGen

Serving LLMs on heterogeneous decentralized clusters.

Language:PythonLicense:Apache-2.0Stargazers:11Issues:0Issues:0

any-precision-llm

[ICML 2024 Oral] Any-Precision LLM: Low-Cost Deployment of Multiple, Different-Sized LLMs

Language:PythonLicense:MITStargazers:33Issues:0Issues:0

AdaQP

Adaptive Message Quantization and Parallelization for Distributed Full-graph GNN Training

Language:PythonLicense:MITStargazers:18Issues:0Issues:0

H2O

[NeurIPS'23] H2O: Heavy-Hitter Oracle for Efficient Generative Inference of Large Language Models.

Language:PythonStargazers:322Issues:0Issues:0
Language:PythonLicense:MITStargazers:8Issues:0Issues:0

ML-Papers-Explained

Explanation to key concepts in ML

Stargazers:6794Issues:0Issues:0

ML-Papers-of-the-Week

🔥Highlighting the top ML papers every week.

Stargazers:9418Issues:0Issues:0

llm-awq

[MLSys 2024 Best Paper Award] AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration

Language:PythonLicense:MITStargazers:2142Issues:0Issues:0
Language:RustStargazers:6Issues:0Issues:0

Temporal_Fusion_Transform

Pytorch Implementation of Google's TFT

Language:Jupyter NotebookStargazers:234Issues:0Issues:0

tft-pytorch

Pytorch Temporal Fusion Transformers

Language:PythonStargazers:3Issues:0Issues:0

tft-torch

A Python library that implements ×´Temporal Fusion Transformers for Interpretable Multi-horizon Time Series Forecasting×´

Language:PythonLicense:MITStargazers:101Issues:0Issues:0

QSync

Official resporitory for "IPDPS' 24 QSync: Quantization-Minimized Synchronous Distributed Training Across Hybrid Devices".

Language:C++License:MITStargazers:19Issues:0Issues:0

llm-papers

List of Large Lanugage Model Papers

License:Apache-2.0Stargazers:50Issues:0Issues:0

Megatron-LLaMA

Best practice for training LLaMA models in Megatron-LM

Language:PythonLicense:NOASSERTIONStargazers:570Issues:0Issues:0
Language:PythonStargazers:9Issues:0Issues:0