dachao .Wang's starred repositories
smoothquant
[ICML 2023] SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models
Megatron-LLaMA
Best practice for training LLaMA models in Megatron-LM
Megatron-LM
Ongoing research training transformer models at scale
HighPerformanceComputing
Class of High Performance Computing taken at U.T.P 2017
llm_training_handbook
An open collection of methodologies to help with successful training of large language models.
LLMDataHub
A quick guide (especially) for trending instruction finetuning datasets
flash-attention
Fast and memory-efficient exact attention
CaptuocrToy
A tool to capture screenshot and recognize text by online ocr apis
full_stack
必要的计算机科学及软件开发知识
inference
Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with any open-source language models, speech recognition models, and multimodal models, whether in the cloud, on-premises, or even on your laptop.
Awesome-LLM-Inference
📖A curated list of Awesome LLM Inference Paper with codes, TensorRT-LLM, vLLM, streaming-llm, AWQ, SmoothQuant, WINT8/4, Continuous Batching, FlashAttention, PagedAttention etc.
personal_chatgpt
personal chatgpt
FightingCV-Paper-Reading
⭐⭐⭐FightingCV Paper Reading, which helps you understand the most advanced research work in an easier way 🍀 🍀 🍀
External-Attention-pytorch
🍀 Pytorch implementation of various Attention Mechanisms, MLP, Re-parameter, Convolution, which is helpful to further understand papers.⭐⭐⭐
CVPR2024-Papers-with-Code
CVPR 2024 论文和开源项目合集