xianyuanliu / CTAN

An adversarial channel-temporal attention network on UDA for first-person action recognition.

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

First-Person Video Domain Adaptation with Multi-Scene Cross-Site Datasets and Attention-Based Methods

arXiv

Introduction

This repository contains the PyTorch implementation of the CTAN framework on UDA for first-person action recognition. CTAN is an adversarial channel-temporal attention network comprised of an I3D backbone, channel- and temporal-wise attention modules, and adversarial UDA modules trained on EPIC-8 [1], ADL-7, and GTEA_KITCHEN-6 datasets.

Framework

framework

Datasets

Prepare the datasets (EPIC-8 [1], ADL-7, and GTEA KITCHEN-6) according to the instructions below and place them in the ./data directory.

EPIC-8

Download RGB frames from participants P01, P08 and P22 of the EPIC-KITCHENS-55 dataset, using official download script here.

ADL-7 and GTEA_KITCHEN-6

Follow the instructions in EgoAction.

Usage

Clone the repo

git clone https://github.com/XianyuanLiu/CTAN.git

Install the required packages

conda create --n ctan; conda activate ctan  (Optional)

# install pytorch
pip install torch==1.9.1+cu111 torchvision==0.10.1+cu111 torchaudio==0.9.1 -f https://download.pytorch.org/whl/torch_stable.html

pip install -r requirements.txt

Run CTAN on EPIC-8 [1], ADL-7 and GTEA_KITCHEN-6 respectively (For more domains, please edit the yaml files in ./configs)

python main.py --cfg configs/EPIC-D12D2.yaml

python main.py --cfg configs/ADL-D12D2.yaml

python main.py --cfg configs/G2K.yaml

Citation

@article{liu2023first,
  title={First-person video domain adaptation with multi-scene cross-site datasets and attention-based methods},
  author={Liu, Xianyuan and Zhou, Shuo and Lei, Tao and Jiang, Ping and Chen, Zhixiang and Lu, Haiping},
  journal={IEEE Transactions on Circuits and Systems for Video Technology},
  year={2023},
  publisher={IEEE},
  doi={10.1109/TCSVT.2023.3281671},
  html={https://ieeexplore.ieee.org/document/10139790},
}

Acknowledgement

This implementation is based on early works [1] and [2].

References

[1] Munro Jonathan and Damen Dima (2020). Multi-modal Domain Adaptation for Fine-grained Action Recognition. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR)
[2] Haiping Lu, Xianyuan Liu, Shuo Zhou, Robert Turner, Peizhen Bai, ... & Hao Xu (2022). PyKale: Knowledge-Aware Machine Learning from Multiple Sources in Python. In Proceedings of the 31st ACM International Conference on Information and Knowledge Management (CIKM).

About

An adversarial channel-temporal attention network on UDA for first-person action recognition.

License:MIT License


Languages

Language:Python 100.0%