Do you consider adding rwkv
chaodreaming opened this issue · comments
chaodreaming commented
Do you consider adding rwkv
Phil Wang commented
no, i don't really believe in that approach anymore
A concise but complete full-attention transformer with a set of promising experimental features from various papers
chaodreaming opened this issue · comments
no, i don't really believe in that approach anymore