mit-han-lab / lite-transformer

[ICLR 2020] Lite Transformer with Long-Short Range Attention

Home Page:https://arxiv.org/abs/2004.11886

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Please share your quantization, quantization+pruning checkpoints

kishorepv opened this issue · comments

Hi,

Can you please share the trained checkpoints for the quantized and quantized+pruned models (shown in this plot - https://github.com/mit-han-lab/lite-transformer#further-compress-transformer-by-182x)?

I am interested in testing it for the translation and the summarization tasks. I would appreciate it if you can share those checkpoints.

Thank you