BeyondYourself / CRCNN-Action

Context-aware RCNN: a Baseline for Action Detection in Videos

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Context-aware RCNNs: a Baseline for Action Detection in Videos

Source code for the following paper(arXiv link):

Context-aware RCNNs: a Baseline for Action Detection in Videos
Jianchao Wu, Zhanghui Kuang, Limin Wang, Wayne Zhang, Gangshan Wu
in ECCV 2020

Our implementation is based on Video-long-term-feature-banks.

Prepare dataset

Please follow LFB on how to prepare AVA dataset.

Prepare environment

Please follow LFB on how to prepare Caffe2 environment.

Download pre-trained weights

Please download R50-I3D-NL, and put it in [code root]/pretrained_weights folder.

Train a baseline model without scene feature and long-term feature

Run:

bash train_baseline.sh configs/avabox_r50_baseline_32x2_scale1_5.yaml

Train a model with scene feature

Run:

bash train_baseline.sh configs/avabox_r50_baseline_16x4_scale1_5_withScene.yaml

Train a model with scene feature and long-term feature

Stage1. Train a baseline model that will be used to infer LFB:

bash train_baseline.sh configs/avabox_r50_baseline_16x4_scale1_5.yaml

Stage2. Train a model with scene feature and LFB:

bash train_lfb.sh configs/avabox_r50_lfb_win60_L3_16x4_withScene.yaml [path to baseline model weight from step1]

About

Context-aware RCNN: a Baseline for Action Detection in Videos


Languages

Language:Python 96.8%Language:Shell 1.8%Language:C++ 0.6%Language:Cuda 0.5%Language:CMake 0.2%