notnitsuj / OSTrack

Joint Feature Learning and Relation Modeling for Tracking: A One-Stream Framework

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

OSTrack

The official implementation for the paper Joint Feature Learning and Relation Modeling for Tracking: A One-Stream Framework

PWC

PWC

PWC

Framework

โ˜€๏ธ Highlights

๐ŸŒŸ New One-stream Tracking Framework

OSTrack is a simple, neat, high-performance one-stream tracking framework for joint feature learning and relational modeling based on self-attention operators. Without any additional temporal information, OSTrack achieves SOTA performance on multiple benchmarks. OSTrack can serve as a strong baseline for further research.

Tracker GOT-10K (AO) LaSOT (AUC) TrackingNet (AUC) UAV123(AUC)
OSTrack-384 73.7 71.1 83.9 70.7
OSTrack-256 71.0 69.1 83.1 68.3

๐ŸŒŸ Fast Training

OSTrack-256 can be trained in ~23 hours with 4*2080Ti (11GB of memory per GPU), which is much faster than recent SOTA transformer-based trackers. The fast training speed comes from:

  1. While previous Siamese-stype trackers required separate feeding of the template and search region into the backbone at each iteration of training, OSTrack directly combines the template and search region. The tight and highly parallelized structure results in improved training and inference speed.

  2. The proposed early candidate elimination (ECE) module significantly reduce memory and time consumption.

  3. Pretrained Transformer weights enables faster convergence.

๐ŸŒŸ Good performance-speed trade-off

speed_vs_performance

Install the environment

Option1: Use the Anaconda

conda create -n ostrack python=3.8
conda activate ostrack
bash install.sh

Option2: Use the docker file

We provide the full docker file here.

Set project paths

Run the following command to set paths for this project

python tracking/create_default_local_file.py --workspace_dir . --data_dir ./data --save_dir ./output

After running this command, you can also modify paths by editing these two files

lib/train/admin/local.py  # paths about training
lib/test/evaluation/local.py  # paths about testing

Data Preparation

Put the tracking datasets in ./data. It should look like:

${PROJECT_ROOT}
 -- data
     -- lasot
         |-- airplane
         |-- basketball
         |-- bear
         ...
     -- got10k
         |-- test
         |-- train
         |-- val
     -- coco
         |-- annotations
         |-- images
     -- trackingnet
         |-- TRAIN_0
         |-- TRAIN_1
         ...
         |-- TRAIN_11
         |-- TEST

Training

Download pre-trained MAE ViT-Base weights and put it under $PROJECT_ROOT$/pretrained_models (different pretrained models can also be used, see MAE for more details).

python tracking/train.py --script ostrack --config vitb_256_mae_ce_32x4_ep300 --save_dir ./output --mode multiple --nproc_per_node 4 --use_wandb 1

Replace --config with the desired model config under experiments/ostrack. We use wandb to record detailed training logs, in case you don't want to use wandb, set --use_wandb 0.

Evaluation

Download the model weights from Google Drive

Put the downloaded weights on $PROJECT_ROOT$/output/checkpoints/train/ostrack

Change the corresponding values of lib/test/evaluation/local.py to the actual benchmark saving paths

Some testing examples:

  • LaSOT or other off-line evaluated benchmarks (modify --dataset correspondingly)
python tracking/test.py ostrack vitb_384_mae_ce_32x4_ep300 --dataset lasot --threads 16 --num_gpus 4
python tracking/analysis_results.py # need to modify tracker configs and names
  • GOT10K-test
python tracking/test.py ostrack vitb_384_mae_ce_32x4_ep300 --dataset got10k_test --threads 16 --num_gpus 4
python lib/test/utils/transform_got10k.py --tracker_name ostrack --cfg_name ostrack384_got10k_elimination_ep100
  • TrackingNet
python tracking/test.py ostrack vitb_384_mae_ce_32x4_ep300 --dataset trackingnet --threads 16 --num_gpus 4
python lib/test/utils/transform_trackingnet.py --tracker_name ostrack --cfg_name ostrack384_elimination_ep300

Visualization or Debug

Visdom is used for visualization.

  1. Alive visdom in the sever by running visdom:

  2. Simply set --debug 1 during inference for visualization, e.g.:

python tracking/test.py ostrack ostrack384_elimination_ep300 --dataset vot22 --threads 1 --num_gpus 1 --debug 1
  1. Open http://localhost:8097 in your browser (remember to change the ip address and port according to the actual situation).

  2. Then you can visualize the candidate elimination process.

ECE_vis

Test FLOPs, and Speed

Note: The speeds reported in our paper were tested on a single RTX2080Ti GPU.

# Profiling vitb_256_mae_ce_32x4_ep300
python tracking/profile_model.py --script ostrack --config vitb_256_mae_ce_32x4_ep300
# Profiling vitb_384_mae_ce_32x4_ep300
python tracking/profile_model.py --script ostrack --config vitb_384_mae_ce_32x4_ep300

Acknowledgments

  • Thanks for the STARK and PyTracking library, which helps us to quickly implement our ideas.
  • We use the implementation of the ViT from the Timm repo.

Citation

If our work is useful for your research, please consider cite:

@article{ye2022ostrack,
  title={Joint Feature Learning and Relation Modeling for Tracking: A One-Stream Framework},
  author={Ye, Botao and Chang, Hong and Ma, Bingpeng and Shan, Shiguang},
  journal={arXiv preprint arXiv:2203.11991},
  year={2022}
}

About

Joint Feature Learning and Relation Modeling for Tracking: A One-Stream Framework

License:MIT License


Languages

Language:Python 98.5%Language:Dockerfile 0.7%Language:Jupyter Notebook 0.4%Language:Shell 0.4%