GeWu-Lab / LFAV

Towards Long Form Audio-visual Video Understanding

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Towards Long Form Audio-visual Video Understanding (ACM TOMM 2024).

Project Page

Paper

Dataset & Features

YouTube ID

The dataset is collected from YouTube, you can find the ID of each video in annotation files.

Features

We use VGGish to extract audio features, use ResNet18 and R(2+1)D-18 to extract visual features.

VGGish feature: Google Drive, Baidu Drive (pwd: lfav), (~662M).

ResNet18 feature: Google Drive, Baidu Drive (pwd: lfav), (~2.6G).

R(2+1)D-18feature: Google Drive, Baidu Drive (pwd: lfav), (~2.6G).

Annotations

Label files are in the folder LFAV_dataset.

training set

# LFAV training set annotations
cd LFAV_dataset
cd ./train
train_audio_weakly.csv: video-level audio annotaions of training set
train_visual_weakly.csv: video-level visual annotaions of training set
train_weakly.csv: video-level annotations (union of video-level audio annotations and visual annotations) of training set

validation set

# LFAV validation set annotations
cd LFAV_dataset
cd ./val
val_audio_weakly.csv: video-level audio annotaions of validation set
val_visual_weakly.csv: video-level visual annotaions of validation set
val_weakly_av.csv: video-level annotations (union of video-level audio annotations and visual annotations) of validation set
val_audio.csv: event-level audio annotaions of validation set
val_visual.csv: event-level visual annotaions of validation set

testing set

# LFAV testing set annotations
cd LFAV_dataset
cd ./test
test_audio_weakly.csv: video-level audio annotaions of testing set
test_visual_weakly.csv: video-level visual annotaions of testing set
test_weakly_av.csv: video-level annotations (union of video-level audio annotations and visual annotations) of testing set
test_audio.csv: event-level audio annotaions of testing set
test_visual.csv: event-level visual annotaions of testing set

Train and test

Source code is in the folder src.

The script of training all three phases is in:

src/scripts/train_s3.sh

If you want to train one or two phases, just edit the arg "num_stages" to 1 or 2.

The script of testing all three phases is in:

src/scripts/test_s3.sh

We also provide our trained weights of the complete method (three phases): Google Drive, Baidu Drive (pwd: lfav).

Publication(s)

If you find our work useful in your research, please cite our paper.

@article{hou2023towards,
  title={Towards Long Form Audio-visual Video Understanding},
  author={Hou, Wenxuan and Li, Guangyao and Tian, Yapeng and Hu, Di},
  journal={ACM Transactions on Multimedia Computing, Communications and Applications},
  year={2023},
  publisher={ACM New York, NY}
}

Acknowledgement

This research was supported by National Natural Science Foundation of China (NO.62106272), and Public Computing Cloud, Renmin University of China.

The source code referenced AVVP-ECCV20.

License

This project is released under the CC BY-NC 4.0 License.

About

Towards Long Form Audio-visual Video Understanding

License:MIT License


Languages

Language:Python 96.9%Language:Shell 3.1%