RobertCsordas / moe_attention

Official repository for the paper "SwitchHead: Accelerating Transformers with Mixture-of-Experts Attention"

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

RobertCsordas/moe_attention Issues