Mingcan Xiang's starred repositories
PowerInfer
High-speed Large Language Model Serving on PCs with Consumer-grade GPUs
Awesome-LLM-Inference
📖A curated list of Awesome LLM Inference Paper with codes, TensorRT-LLM, vLLM, streaming-llm, AWQ, SmoothQuant, WINT8/4, Continuous Batching, FlashAttention, PagedAttention etc.
flash-attention
Fast and memory-efficient exact attention
autoLiterature
autoLiterature是一个基于Python的自动文献管理命令行工具
Awesome-Multimodal-Large-Language-Models
:sparkles::sparkles:Latest Papers and Datasets on Multimodal Large Language Models, and Their Evaluation.
mPLUG-DocOwl
mPLUG-DocOwl: Modularized Multimodal Large Language Model for Document Understanding
Transformers-Tutorials
This repository contains demos I made with the Transformers library by HuggingFace.
transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
text-to-text-transfer-transformer
Code for the paper "Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer"
awesome-mixture-of-experts
A collection of AWESOME things about mixture-of-experts
Awesome-Mixture-of-Experts-Papers
A curated reading list of research in Mixture-of-Experts(MoE).
RWKV-LM
RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.
flops-counter.pytorch
Flops counter for convolutional networks in pytorch framework
ColossalAI
Making large AI models cheaper, faster and more accessible
Efficient-Deep-Learning
Collection of recent methods on (deep) neural network compression and acceleration.