liuzhuang13 / rethinking-network-pruning

Rethinking the Value of Network Pruning (Pytorch)

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Rethinking the Value of Network Pruning

This repository contains the code for reproducing the results, and trained ImageNet models, in the following arXiv paper:

Rethinking the Value of Network Pruning

Zhuang Liu*, Mingjie Sun*, Tinghui Zhou, Gao Huang, Trevor Darrell (* equal contribution).

This paper shows that training the small pruned model from scratch can almost always achieve comparable or higher level of accuracy than the model obtained from the typical "training, pruning and fine-tuning procedure". Our results suggest:

  1. Training a large, over-parameterized model is not necessary to obtain an efficient final model.
  2. Learned "important" weights of the large model are not necessarily useful for the small pruned model.
  3. The pruned architecture itself, rather than a set of inherited "important" weights, is what leads to the efficiency benefit in the final model, which suggests that some pruning algorithms could be seen as performing network architecture search.

Citation

@article{liu2018rethinking,
  title={Rethinking the Value of Network Pruning},
  author={Zhuang Liu and Mingjie Sun and Tinghui Zhou and Gao Huang and Trevor Darrell},
  journal={arXiv preprint arXiv:1810.05270},
  year={2018}
}

Implementation

We evaluated the following six pruning methods:

  1. L1-norm based channel pruning
  2. ThiNet
  3. Regression based feature reconstruction
  4. Network Slimming
  5. Sparse Structure Selection
  6. Non-structured weight-level pruning

For CIFAR, our code is based on pytorch-classification and network-slimming. For ImageNet, we use the official Pytorch ImageNet training code. The instructions and models are in each subfolder.

Contact

sunmj15 at gmail.com
liuzhuangthu at gmail.com

About

Rethinking the Value of Network Pruning (Pytorch)

License:MIT License


Languages

Language:Python 100.0%