Yu Zhang's starred repositories
awesome-mixture-of-experts
A collection of AWESOME things about mixture-of-experts
review-2023
二〇二三年的年终总结都写好了吗?
zero-bubble-pipeline-parallelism
Zero Bubble Pipeline Parallelism
SpeculativeDecodingPapers
📰 Must-read papers and blogs on Speculative Decoding ⚡️
lightning-attention
Lightning Attention-2: A Free Lunch for Handling Unlimited Sequence Lengths in Large Language Models
accelerated-scan
Accelerated First Order Parallel Associative Scan
moe_attention
Official repository for the paper "SwitchHead: Accelerating Transformers with Mixture-of-Experts Attention"
triton-autodiff
Experiment of using Tangent to autodiff triton
top_k_attention
The accompanying code for "Memory-efficient Transformers via Top-k Attention" (Ankit Gupta, Guy Dar, Shaya Goodman, David Ciprut, Jonathan Berant. SustaiNLP 2021).
llm-misinformation-survey
Paper list for the survey "Combating Misinformation in the Age of LLMs: Opportunities and Challenges" and the initiative "LLMs Meet Misinformation"
why-weight-decay
Why Do We Need Weight Decay in Modern Deep Learning? [arXiv, Oct 2023]
Highway-Transformer
[ACL‘20] Highway Transformer: A Gated Transformer.
Awesome-Simultaneous-Translation
Paper list of simultaneous translation / streaming translation, including text-to-text machine translation and speech-to-text translation.