bobleono / PatchDCT

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

PatchDCT: Patch Refinement for High Quality Instance Segmentation(ICLR 2023)

[PatchDCT: Patch Refinement for High Quality Instance Segmentation] Qinrou Wen, Jirui Yang, Xue Yang, Kewei Liang

arXiv preprint(arXiv:2302.02693)

In this repository, we release code for PatchDCT in Detectron2.

Contributions

  • PatchDCT is the fist compressed vector based multi-stage refinement framework.
  • By using a classifier to refine foreground and background patches, and predicting an informative low-dimensional DCT vector for each mixed patch, PatchDCT generates high-resolution masks with fine boundaries and low computational cost.

Installation

Requirements

  • PyTorch ≥ 1.8

This implementation is based on detectron2. Please refer to INSTALL.md. for installation and dataset preparation.

Usage

The codes of this project is on projects/PatchDCT/

Train with multiple GPUs

cd ./projects/PatchDCT/
./train.sh

Testing

cd ./projects/PatchDCT/
./test.sh

Speed Testing

cd ./projects/PatchDCT/
./test_speed.sh

Upper Bound of Model Performance(Table 1 in the paper)

cd ./projects/PatchDCT/
./test_up.sh

For Swin-B backbone, use train_net_swinb.py instead of train_net.py

Model ZOO

Trained models on COCO

Model Backbone Schedule Multi-scale training FPS AP (val) Link
PatchDCT R50 1x Yes 12.3 37.2 download
PatchDCT R101 3x Yes 11.8 40.5 download
PatchDCT RX101 3x Yes 11.7 41.8 download
PatchDCT SwinB 3x Yes 7.3 46.1 download

Trained models on Cityscapes

Model Data Backbone Schedule Multi-scale training AP (val) Link
PatchDCT Fine-Only R50 1x Yes 38.2 download
PatchDCT COCO Pretrain+Fine R50 1x Yes 40.3 download

Notes

  • We observe about 0.2 AP noise in COCO.
  • The inference time is measured on NVIDIA A100 with batchsize=1.
  • Lvis 0.5 is used for evaluation.

Contributing to the project

Any pull requests or issues are welcome.

If there is any problem with this project, please contact Qinrou Wen.

Citations

Please consider citing our papers in your publications if the project helps your research.

License

  • MIT License.

About

License:Apache License 2.0


Languages

Language:Python 94.1%Language:Cuda 3.1%Language:C++ 2.2%Language:Shell 0.4%Language:Dockerfile 0.1%Language:CMake 0.0%