hustzxd / LSQuantization

The PyTorch implementation of Learned Step size Quantization (LSQ) in ICLR2020 (unofficial)

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

LSQuantization

The PyTorch implementation of Learned Step size Quantization (LSQ) in ICLR2020 (unofficial)


The related project with training code: https://github.com/hustzxd/EfficientPyTorch (sorry for late.)

The project is working in progress, and experimental results on ImageNet are not as good as shown in the paper.

ImageNet

LSQ fp32 w4a4 w3a3 w2a2 w8a8(1epoch, quantize data)
AlexNet 56.55, 79.09 56.96, 79.46 55.31, 78.59 51.18, 75.38
ResNet18 69.76, 89.08 70.26, 89.34 69.45, 88.85 69.68 88.92

Hyper-parameter

Hyper-parameter LR LR-scheduler epochs batch-size wd
AlexNet-w4a4 0.01 CosineAnnealingLR 90 2048 1e-4
ResNet18-w4a4 0.01 CosineAnnealingLR 90 512 1e-4

Experimental Results

====VGGsmall + Cifar10=======

VGGsmall
fp32 93.34
w4a4 94.26
w3a3 93.89
w2a2 93.42

About

The PyTorch implementation of Learned Step size Quantization (LSQ) in ICLR2020 (unofficial)


Languages

Language:Jupyter Notebook 82.0%Language:Python 18.0%