KD |
Distilling the Knowledge in a Neural Network |
|
√ |
FitNet |
Fitnets: hints for thin deep nets |
|
|
AT |
Paying More Attention to Attention: Improving the Performance of Convolutional Neural Networks via Attention Transfer |
|
|
SP |
Similarity-Preserving Knowledge Distillation |
|
|
CC |
Correlation Congruence for Knowledge Distillation |
|
|
VID |
Variational Information Distillation for Knowledge Transfer |
|
|
RKD |
Relational Knowledge Distillation |
|
|
PKT |
Probabilistic Knowledge Transfer for deep representation learning |
|
|
AB |
Knowledge Transfer via Distillation of Activation Boundaries Formed by Hidden Neurons |
|
|
FT |
Paraphrasing Complex Network: Network Compression via Factor Transfer |
|
|
FSP |
A Gift from Knowledge Distillation: Fast Optimization, Network Minimization and Transfer Learning |
教师、学生网络对应的中间特征,其通道数必须相同 |
√ |
NST |
Like what you like: knowledge distill via neuron selectivity transfer |
|
|
CRD |
Contrastive Representation Distillation ICLR 2020 |
|
|