3neutronstar / knowledge-distillation

Knowledge Distillation

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

knowledge-distillation

Knowledge Distillation

How To Use

  • Run the basic model
    python run.py train --model [model]
  • Run Self Knowledge Distillation(cs-kd,pearson)
    python run.py train_selfkd --model [model] --custom_loss [self-kd name]
  • Run Offline Knowledge Distillation(softtarget)
    python run.py train_offkd --model [model] --pretrained_model [teacher model]
  • Run EnsembleKnowledge Distillation(deepmutuallearning)
    python run.py train_offkd --model [model] --pretrained_model [teacher model]

Performance

  • Cifar100, (Temperature: 2.0 in SoftTarget)
Type Teacher Student Best Eval Accuracy(%)
Baseline None ConvNet 56.31
SoftTarget(KL) ResNet20 ConvNet 61.21
DeepMutualLearning None ConvNet 56.52
  • Cifar10, (Temperature: 2.0 in SoftTarget)
Type Teacher Student Best Eval Accuracy(%)
Baseline None ConvNet 83.52
Baseline None ResNet20 90.86
DeepMutualLearning None ConvNet 83.30
DeepMutualLearning None ResNet20 91.14
  • FASHIONMNIST
Type Teacher Student Best Eval Accuracy(%)
Baseline None LeNet5 90.49
DeepMutualLearning None LeNet5 90.61
CS-KD None LeNet5 91.45

About

Knowledge Distillation


Languages

Language:Python 99.4%Language:Shell 0.6%