Gyeongchan-Yun's starred repositories
Awesome_LLM_System-PaperList
Since the emergence of chatGPT in 2022, the acceleration of Large Language Model has become increasingly important. Here is a list of papers on accelerating LLMs, currently focusing mainly on inference acceleration, and related works will be gradually added in the future. Welcome contributions!
Megatron-Kwai
Ongoing research training transformer models at scale
Megatron-LM
Artifact for DynaPipe: Optimizing Multi-task Training through Dynamic Pipelines
python-patterns
A collection of design patterns/idioms in Python
awesome-AI-system
paper and its code for AI System
zero-bubble-pipeline-parallelism
Zero Bubble Pipeline Parallelism
optimizing-multitask-training-through-dynamic-pipelines
Official repository for the paper DynaPipe: Optimizing Multi-task Training through Dynamic Pipelines
Optimus-CC
[ASPLOS'23] Optimus-CC: Efficient Large NLP Model Training with 3D Parallelism Aware Communication Compression
Hetu-Galvatron
Galvatron is an automatic distributed training system designed for Transformer models, including Large Language Models (LLMs). If you have any interests, please visit/star/fork https://github.com/PKU-DAIR/Hetu-Galvatron
any-precision-llm
[ICML 2024 Oral] Any-Precision LLM: Low-Cost Deployment of Multiple, Different-Sized LLMs
ML-Papers-Explained
Explanation to key concepts in ML
ML-Papers-of-the-Week
🔥Highlighting the top ML papers every week.
Temporal_Fusion_Transform
Pytorch Implementation of Google's TFT
tft-pytorch
Pytorch Temporal Fusion Transformers
llm-papers
List of Large Lanugage Model Papers
Megatron-LLaMA
Best practice for training LLaMA models in Megatron-LM