BlueRum's repositories

awesome-Auto-Parallelism

A baseline repository of Auto-Parallelism in Training Neural Networks

Language:PythonStargazers:0Issues:0Issues:0

awesome-RLHF

collecting RLHF papers

License:Apache-2.0Stargazers:0Issues:1Issues:0

binary-bert

Pretrained language model and its related optimization techniques developed by Huawei Noah's Ark Lab.

Language:PythonStargazers:0Issues:0Issues:0

diffusers

🤗 Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

EnergonAI

Large-scale model inference.

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

Best-README-Template

An awesome README template to jumpstart your projects!

License:MITStargazers:0Issues:0Issues:0

binary-quantization-Meta

Code repo for the paper BiT Robustly Binarized Multi-distilled Transformer

Language:PythonLicense:NOASSERTIONStargazers:0Issues:0Issues:0

bitsandbytes

8-bit CUDA functions for PyTorch

Language:PythonLicense:MITStargazers:0Issues:0Issues:0
Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

ColossalAI

Colossal-AI: A Unified Deep Learning System for Big Model Era

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0
Language:PythonStargazers:0Issues:1Issues:0

ColossalAI-Examples

Examples of training models with hybrid parallelism using ColossalAI

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

DeepSpeed

DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

License:MITStargazers:0Issues:0Issues:0

FQ-ViT

[IJCAI 2022] FQ-ViT: Post-Training Quantization for Fully Quantized Vision Transformer

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

InfiAgent.github.io

InfiAgent website

Language:JavaScriptStargazers:0Issues:0Issues:0
Language:PythonStargazers:0Issues:1Issues:0
Language:PythonStargazers:0Issues:1Issues:0

lightllm

LightLLM is a Python-based LLM (Large Language Model) inference and serving framework, notable for its lightweight design, easy scalability, and high-speed performance.

License:Apache-2.0Stargazers:0Issues:0Issues:0

lightseq

LightSeq: A High Performance Library for Sequence Processing and Generation

Language:C++License:NOASSERTIONStargazers:0Issues:0Issues:0

MNN

MNN is a blazing fast, lightweight deep learning framework, battle-tested by business-critical use cases in Alibaba

Language:C++Stargazers:0Issues:0Issues:0

model-quantization

A list of papers, docs, codes about model quantization. This repo is aimed to provide the info for model quantization research, we are continuously improving the project. Welcome to PR the works (papers, repositories) that are missed by the repo.

Stargazers:0Issues:0Issues:0

Nonuniform-to-Uniform-Quantization

Nonuniform-to-Uniform Quantization: Towards Accurate Quantization via Generalized Straight-Through Estimation. In CVPR 2022.

Language:PythonStargazers:0Issues:0Issues:0

parallelformers

Parallelformers: An Efficient Model Parallelization Toolkit for Deployment

Language:PythonLicense:Apache-2.0Stargazers:0Issues:0Issues:0

prm800k

800,000 step-level correctness labels on LLM solutions to MATH problems

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

smoothquant

SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models

Language:PythonLicense:MITStargazers:0Issues:0Issues:0

stable-baselines3

PyTorch version of Stable Baselines, reliable implementations of reinforcement learning algorithms.

Language:PythonLicense:MITStargazers:0Issues:0Issues:0
Language:C++Stargazers:0Issues:0Issues:0

transformers

🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.

License:Apache-2.0Stargazers:0Issues:0Issues:0

trlx

A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)

Language:PythonLicense:MITStargazers:0Issues:0Issues:0