Tebmer / Awesome-Knowledge-Distillation-of-LLMs

This repository collects papers for "A Survey on Knowledge Distillation of Large Language Models". We break down KD into Knowledge Elicitation and Distillation Algorithms, and explore the Skill & Vertical Distillation of LLMs.

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

[One latest paper] Rethinking Kullback-Leibler Divergence in Knowledge Distillation for Large Language Models

wutaiqiang opened this issue · comments

Nice work!
One missing related work:
Rethinking Kullback-Leibler Divergence in Knowledge Distillation for Large Language Models
https://arxiv.org/abs/2404.02657

Great work!

Thanks. We have added it and will update paper in a future version :)

This paper is under review in COLM rather than accepted.

Also, this paper rethinks the FKL and RKL in logit-based distillation and proposes AKL.

One blog: https://zhuanlan.zhihu.com/p/690748958

Thanks for your effort~

Thanks for your reminders. We have corrected it.

We added this paper to the "Feature" and "Divergence and Similarity" categories.