Zian(Andy) Zheng's starred repositories
alignment-handbook
Robust recipes to align language models with human and AI preferences
Skywork-MoE
Skywork-MoE: A Deep Dive into Training Techniques for Mixture-of-Experts Language Models
PowerInfer
High-speed Large Language Model Serving on PCs with Consumer-grade GPUs
pytorch-learning
learning notes when learning the source code of pytorch
compendium
Dota 2 replay knowledge in book form.
Megatron-DeepSpeed
Ongoing research training transformer language models at scale, including: BERT & GPT-2
dota2-clarity
Custom console scripts for Dota 2.
DeepLearningSystem
AI Infra主要是指AI的基础建设,包括AI芯片、AI编译器、AI推理和训练框架等AI全栈底层技术。
llm-colosseum
Benchmark LLMs by fighting in Street Fighter 3! The new way to evaluate the quality of an LLM
llm-compressive
Longitudinal Evaluation of LLMs via Data Compression
Awesome-LLM-Inference
📖A curated list of Awesome LLM Inference Paper with codes, TensorRT-LLM, vLLM, streaming-llm, AWQ, SmoothQuant, WINT8/4, Continuous Batching, FlashAttention, PagedAttention etc.
Neural-Network-Parameter-Diffusion
We introduce a novel approach for parameter generation, named neural network parameter diffusion (p-diff), which employs a standard latent diffusion model to synthesize a new set of parameters
G_VBSM_Dataset_Condensation
[CVPR2024 highlight] Generalized Large-Scale Data Condensation via Various Backbone and Statistical Matching (G-VBSM)
onboarding
Onboarding guide to Jimmy Lin's research group at the University of Waterloo