AllenXuuu / DCR

Official implementation of our CVPR'22 paper.

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

DCR

This repo contains the official implementation of paper

Learning to Anticipate Future with Dynamic Context Removal
Xinyu Xu, Yong-Lu Li, Cewu Lu.

In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition 2022.

[arxiv] [code] [model]


Data Preparation

We reorganize the annotation files of four datasets[1-4] in data folder.
You need to download pre-extracted feature into data/feature folder.
TSN feature can be downloaded from Link [5].
irCSN-152 feature can be downloaded from Link [6].
We provide a stronger TSM backbone, feature can be be downloaded from Link.

Packages

We conduct experiments in the following environment

python == 3.9

torch == 1.9
torchvision == 0.10.0
apex == 0.1.0
tensorboardX
yacs
pyyaml
numpy
prefetch_generator

Evaluation

We release pre-trained models at here.
To test the performance of our model, for example using RGB-TSM backbone on EPIC-KITCHENS-100[1], you can run the following command.

python eval.py --cfg configs/EK100RGBTSM/eval.yaml --resume ./weights/EK100RGBTSM.pt

Here ./weights/EK100RGBTSM.pt is the path to the pre-trained model you downloaded.

To do the late fusion, you need to store the predicted results of each model first, then run the fusion script. For example

python eval_and_extract.py --cfg configs/EK100RGBTSM/eval.yaml --resume ./weights/EK100RGBTSM.pt

python fuse/fuse_EK100.py

The following is expected validation set performace.

EPIC-KITCHENS-100
Method Overall Unseen Tail
RULSTM 14.0 14.1 11.1
ActionBanks 14.7 14.5 11.8
TransAction 16.6 13.8 15.5
AVT 15.9 11.9 14.1
DCR 18.3 14.7 15.8
EPIC-KITCHENS-55
Method Top-1 Top-5
ATSN - 16.3
ED - 25.8
MCE - 26.1
RULSTM 15.3 35.3
FHOI 10.4 25.5
ImagineRNN - 35.6
ActionBanks 15.1 35.6
Ego-OMG 19.2 -
AVT 16.6 37.6
DCR 19.2 41.2
EGTEA GAZE+

We continue to work on EGTEA GAZE+ to improve models.

Method Top-5 Recall@5
DMR 55.7 38.1
ATSN 40.5 31.6
NCE 56.3 43.8
TCN 58.5 47.1
ED 60.2 54.6
RL 62.7 52.2
EL 63.8 55.1
RULSTM 66.4 58.6
DCR(Updated) 67.9 61.3

The EPIC-KITCHENS test set files are at here.

More results can be found in Model Zoo.

Training

Taking the same setting as an example, to reproduce the training process, you can run

python train_order.py --cfg configs/EK100RGBTSM/order.yaml --name order

python train.py --cfg configs/EK100RGBTSM/train.yaml --name train --resume exp/EK100RGBTSM/order/epoch_50.pt

The first line runs our frame order pre-training stage. The model will be stored in exp/EK100RGBTSM/order/epoch_50.pt. The second line reloads the pre-trained model and runs the anticipation training stage.

Only to do the anticipation training from scratch is also possible by

python train.py --cfg configs/EK100RGBTSM/train.yaml --name train 

Citation

If you find our paper or code helpful, please cite our paper.

@inproceedings{xu2022learning,
  title={Learning to Anticipate Future with Dynamic Context Removal },
  author={Xu, Xinyu and Li, Yong-Lu and Lu, Cewu},
  booktitle={CVPR},
  year={2022}
}

Reference

[1] Dima Damen, Hazel Doughty, Giovanni Maria Farinella, Antonino Furnari, Jian Ma, Evangelos Kazakos, Davide Moltisanti, Jonathan Munro, Toby Perrett, Will Price, and Michael Wray. Rescaling egocentric vision: Collection, pipeline and challenges for epic-kitchens-100. International Journal of Computer Vision (IJCV), 2021.

[2] Dima Damen, Hazel Doughty, Giovanni Maria Farinella, Sanja Fidler, Antonino Furnari, Evangelos Kazakos, Davide Moltisanti, Jonathan Munro, Toby Perrett, Will Price, et al. Scaling egocentric vision: The epic-kitchens dataset. In Proceedings of the European Conference on Computer Vision (ECCV), pages 720–736, 2018.

[3] Yin Li, Miao Liu, and James M. Rehg. In the eye of beholder: Joint learning of gaze and actions in first person video. In Proceedings of the European Conference on Computer Vision (ECCV), September 2018.

[4] Sebastian Stein and Stephen J McKenna. Combining embedded accelerometers with computer vision for recognizing food preparation activities. In Proceedings of the 2013 ACM international joint conference on Pervasive and ubiquitous computing, pages 729–738, 2013.

[5] Antonino Furnari and Giovanni Farinella. Rolling-unrolling lstms for action anticipation from first-person video. IEEE transactions on pattern analysis and machine intelligence, 2020.

[6] Rohit Girdhar and Kristen Grauman. Anticipative Video Transformer. In ICCV, 2021.

About

Official implementation of our CVPR'22 paper.

License:Apache License 2.0


Languages

Language:Python 100.0%