BlueRum's repositories
awesome-Auto-Parallelism
A baseline repository of Auto-Parallelism in Training Neural Networks
awesome-RLHF
collecting RLHF papers
binary-bert
Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.
diffusers
🤗 Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch
EnergonAI
Large-scale model inference.
Best-README-Template
An awesome README template to jumpstart your projects!
binary-quantization-Meta
Code repo for the paper BiT Robustly Binarized Multi-distilled Transformer
bitsandbytes
8-bit CUDA functions for PyTorch
ColossalAI
Colossal-AI: A Unified Deep Learning System for Big Model Era
ColossalAI-Examples
Examples of training models with hybrid parallelism using ColossalAI
DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
FQ-ViT
[IJCAI 2022] FQ-ViT: Post-Training Quantization for Fully Quantized Vision Transformer
InfiAgent.github.io
InfiAgent website
lightllm
LightLLM is a Python-based LLM (Large Language Model) inference and serving framework, notable for its lightweight design, easy scalability, and high-speed performance.
lightseq
LightSeq: A High Performance Library for Sequence Processing and Generation
MNN
MNN is a blazing fast, lightweight deep learning framework, battle-tested by business-critical use cases in Alibaba
model-quantization
A list of papers, docs, codes about model quantization. This repo is aimed to provide the info for model quantization research, we are continuously improving the project. Welcome to PR the works (papers, repositories) that are missed by the repo.
Nonuniform-to-Uniform-Quantization
Nonuniform-to-Uniform Quantization: Towards Accurate Quantization via Generalized Straight-Through Estimation. In CVPR 2022.
parallelformers
Parallelformers: An Efficient Model Parallelization Toolkit for Deployment
prm800k
800,000 step-level correctness labels on LLM solutions to MATH problems
smoothquant
SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models
stable-baselines3
PyTorch version of Stable Baselines, reliable implementations of reinforcement learning algorithms.
transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
trlx
A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)