YZCU / REPS

[JAG 2024] REPS: Rotation Equivariant Siamese Network Enhanced by Probability Segmentation for Satellite Video Tracking

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

The official implementation for "REPS: Rotation Equivariant Siamese Network Enhanced by Probability Segmentation for Satellite Video Tracking", International Journal of Applied Earth Observation and Geoinformation (JAG), 2024.

  • Authors: Yuzeng Chen, Yuqi Tang*, Qiangqiang Yuan, Liangpei Zhang
  • Wuhan University and Central South University
  • Download the REPS codes
  • Download the related OOTB dataset on Baidu Cloud Disk (code: OOTB) to your disk, the organized directory looks like:
    --OOTB/
    	|--car_1/
                |--img/
                    |--0001.jpg
                    |--...
                    |--0268.jpg
                |--groundtruth.txt
    	|...
    	|--train_10/
                |--img/
                    |--0001.jpg
                    |--...
                    |--0120.jpg
                |--groundtruth.txt
    	|--anno/
    	    	|--car_1.txt
    	    	|...
    	    	|--train_10.txt
      	|--OOTB.json
    
  • 🧩 Usage: Run the ./tracking/REPS.m
  • Results are saved in ./tracking/results
  • Visual samples

    car image train image

Abstract

Satellite video is an emerging surface observation data that has drawn increasing interest due to its potential in spatiotemporal dynamic analysis. Single object tracking of satellite videos allows the continuous acquisition of the positions and ranges of objects and establishes the correspondences in the video sequence. However, small-sized objects are vulnerable to rotation and non-rigid deformation. Moreover, the horizontal bounding box of most trackers has difficulty in providing accurate semantic representations such as object position, orientation, and spatial distribution. In this article, we propose a unified framework, named rotational equivalent Siamese network enhanced by probability segmentation (REPS), to enhance the tracking accuracy and semantic representations simultaneously. First, to deal with the inconsistency of representations, we design a rotation equivariant (RE) Siamese network architecture to detect the rotation variations of objects right from the start frame, achieving the RE tracking. Second, a pixel-level (PL) refinement is proposed to refine the spatial distribution of objects. In addition, we proposed an adaptive Gaussian fusion that synergizes tracking and segmentation results to obtain compact outputs for satellite object representations. Extensive experiments on satellite videos demonstrate the superiority of the proposed approach. The code will be available at https://github.com/YZCU/REPS.

Overview of REPS

image

Satellite datasets

  • Dataset1 (five videos) image
  • Dataset2: OOTB (110 videos) image

Results

  • Results on Dataset1 image image
  • Success plot for overall and per-dataset on Dataset1 image
  • Overall results on OOTB image
  • Success plot of per-attribute on OOTB image
  • Qualitative results of the top nine trackers on Dataset1 image

Contact

If you have any questions or suggestions, feel free to contact me.
Email: yuzeng_chen@whu.edu.cn

Citation

If you find our work helpful in your research, kindly consider citing it. We appreciate your support!

@article{CHEN2024103741,
title = {REPS: Rotation equivariant Siamese network enhanced by probability segmentation for satellite video tracking},
journal = {International Journal of Applied Earth Observation and Geoinformation},
volume = {128},
pages = {103741},
year = {2024},
issn = {1569-8432},
doi = {https://doi.org/10.1016/j.jag.2024.103741},
url = {https://www.sciencedirect.com/science/article/pii/S1569843224000955},
author = {Yuzeng Chen and Yuqi Tang and Qiangqiang Yuan and Liangpei Zhang},
keywords = {Video satellite, Object tracking, Siamese network, Semantic representation},
}

About

[JAG 2024] REPS: Rotation Equivariant Siamese Network Enhanced by Probability Segmentation for Satellite Video Tracking


Languages

Language:MATLAB 99.8%Language:Shell 0.2%