[ICML'24 Oral] The official code of "DiJiang: Efficient Large Language Models through Compact Kernelization", a novel DCT-based linear attention mechanism.
Home Page:https://arxiv.org/abs/2403.19928
Geek Repo:Geek Repo
Github PK Tool:Github PK Tool