Qinghao Hu's starred repositories
awesome_lists
Awesome Lists for Tenure-Track Assistant Professors and PhD students. (助理教授/博士生生存指南)
flash-linear-attention
Efficient implementations of state-of-the-art linear attention models in Pytorch and Triton
EasyContext
Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.
ring-flash-attention
Ring attention implementation with flash attention
Autonomous-Agents
Autonomous Agents (LLMs) research papers. Updated Daily.
long-context-attention
Sequence Parallel Attention for Long Context LLM Model Training and Inference
superbenchmark
A validation and profiling tool for AI infrastructure
torchsnapshot
A performant, memory-efficient checkpointing library for PyTorch applications, designed with large, complex distributed workloads in mind.