tonysy / deeplab-pytorch

PyTorch implementation of DeepLab v2 (ResNet-101)

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

DeepLab with PyTorch

Unofficial implementation to train DeepLab v2 (ResNet-101) on COCO-Stuff 10k dataset. DeepLab is one of the CNN architectures for semantic image segmentation. COCO Stuff 10k is a semantic segmentation dataset, which includes 10,000 images from 182 thing/stuff classes.

Requirements

Usage

Preparation

Dataset

  1. Download COCO-Stuff 10k dataset and unzip it.
  2. Set the path to the dataset in config/cocostuff.yaml.
cocostuff-10k-v1.1
├── images
│   ├── COCO_train2014_000000000077.jpg
│   └── ...
├── annotations
│   ├── COCO_train2014_000000000077.mat
│   └── ...
└── imageLists
    ├── all.txt
    ├── test.txt
    └── train.txt

Caffemodel

  1. Download init.caffemodel pre-trained on MSCOCO under the directory data/models/deeplab_resnet101/coco_init/.
  2. Convert the caffemodel to pytorch compatible. No need to build the official implementation!
# This generates deeplabv2_resnet101_COCO_init.pth
python convert.py --dataset coco_init

Training

# Training
python train.py --config config/cocostuff.yaml
# Monitoring
tensorboard --logdir runs

See --help for more details.

Default settings

config/cocostuff.yaml

  • All the GPUs visible to the process are used. Please specify the scope with CUDA_VISIBLE_DEVICES=.
  • Stochastic gradient descent (SGD) is used with momentum of 0.9 and initial learning rate of 2.5e-4. Polynomial learning rate decay is employed; the learning rate is multiplied by (1-iter/max_iter)**power at every 10 iterations.
  • Weights are updated 20,000 iterations with mini-batch of 10. The batch is not processed at once due to high occupancy of video memories, instead, gradients of small batches are aggregated, and weight updating is performed at the end (batch_size * iter_size = 10).
  • Input images are randomly scaled by factors ranging from 0.5 to 1.5, zero-padded if needed, and randomly cropped so that the input size is fixed during training (see the example below).
  • Loss is defined as a sum of responses from multi-scale inputs (1x, 0.75x, 0.5x) and element-wise max across the scales. The "unlabeled" class (index -1) is ignored in the loss computation.
  • Moving average loss (average_loss in Caffe) can be monitored in TensorBoard (please specify a log directory, e.g., runs).
  • GPU memory usage is approx. 11.2 GB with the default setting (tested on the single Titan X). You can reduce it with a small batch_size.

Processed image vs. label examples

Data

Evaluation

python eval.py --config config/cocostuff.yaml \
               --model-path <PATH TO MODEL>

You can run with a option --crf. See --help for more details.

Results

After 20k iterations with a mini-batch of 10, no crf

Pixel Accuracy Mean Accuracy Mean IoU Frequency Weighted IoU
20k iterations 64.7% 45.4% 33.9% 50.0%
Official report 65.1% 45.5% 34.4% 50.4%

Demo

From a image

python demo.py --config config/cocostuff.yaml \
               --model-path <PATH TO MODEL> \
               --image-path <PATH TO IMAGE>

From a web camera

python uvcdemo.py --config config/cocostuff.yaml \
                  --model-path <PATH TO MODEL>

Visualize the model on TensorBoard

python draw_model.py

TODO

  • Add DeepLab v3

References

About

PyTorch implementation of DeepLab v2 (ResNet-101)

License:MIT License


Languages

Language:Python 100.0%