GuideWsp / ssdlite-pytorch-mobilenext

A PyTorch implementation of SSDLite on COCO

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Implementation of SSDLite in PyTorch 1.2+

This is the object detection code for our MobileNeXt paper. This repository implements SSDLite, which is presented in MobileNetV2. The implementation is heavily influenced by the SSD project.

Highlights

  • PyTorch 1.2: Support PyTorch 1.2 or higher.
  • Multi-GPU training and inference: We use DistributedDataParallel, you can train or test with arbitrary GPU(s), the training schema will change accordingly.
  • Better results than the original TF version: We achieve an mAP score of 22.3 on COCO, which is slightly better than the original TF version when taking MobileNetV2 as backbone.
  • Implementation of SSDLite with FPN: Support SSDLite with FPN, which highly improves the baseline results.

Experiment Setup

Please refer to INSTALL.md for all the information about installation.

For training on COCO, the batch size is set to 256 and the initial learning rate is set to 0.01. We use 8 GPUs with 12 GB memory to run the experiments with synchronized batch normalization (important). You can also use less GPUs as long as the GPU memory is enough but make sure that the batch size is 256. For example, 4 V100 GPUs are also enough for running.

For training on Pascal VOC, the batch size is set to 24 and the initial learning rate is set to 0.001. We use 4 GPUs with 12 GB memory to run the experiments with the standard batch normalization.

For more implementation details, please refer to the configs in this project.

MODEL ZOO

Pretrained Models:

Networks Links
MobilenetV2 model
MobileNeXt model

COCO:

Backbone Input Size Box AP Model Size Download
MobileNetV2 (SSDLite) 320 22.3 34M model
MobileNeXt (SSDLite) 320 23.3 36M model
MobileNetV2-FPN (SSDLite) 320 23.5 25M [model (coming soon)]
MobileNeXt-FPN (SSDLite) 320 [model (coming soon)]

PASCAL VOC:

Backbone Input Size mAP Model Size Download
VGG16 (SSD) 300 77.7 201MB model
VGG16 (SSDLite) 512 80.7 207MB model
MobilenetV2 (SSDLite) 320 71.8 26MB model
MobileNeXt (SSDLite) 320 72.6 27MB model

Citations

If you use this project in your research, please cite this project.

@inproceedings{daquan2020rethinking,
  title={Rethinking Bottleneck Structure for Efficient Mobile Network Design},
  author={Daquan, Zhou and Hou, Qibin and Chen, Yunpeng and Feng, Jiashi and Yan, Shuicheng},
  booktitle={European conference on computer vision},
  year={2020}
}
@misc{hou2020ssdite-pytorch,
    author = {Qibin Hou},
    title = {{Fast Implementation of SSDLite in PyTorch}},
    year = {2020},
    howpublished = {\url{https://github.com/Andrew-Qibin/ssdlite-pytorch}}
}

About

A PyTorch implementation of SSDLite on COCO

License:MIT License


Languages

Language:Python 99.8%Language:Shell 0.2%