AUST-Hansen / DAIR-V2X-Seq

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

V2X-Seq: A Large-Scale Sequential Dataset for Vehicle-Infrastructure Cooperative Perception and Forecasting, CVPR2023

Table of Contents:

  1. Introduction
  2. News
  3. Dataset Download
  4. Getting Started
  5. Benchmark
  6. TODO List
  7. Citation
  8. Contaction

Introduction

V2X-Seq is the first large-scale, real-world, and sequential V2X dataset, which includes data frames, trajectories, vector maps, and traffic lights captured from natural scenery. V2X-Seq comprises two parts: the sequential perception dataset, which includes more than 15,000 frames captured from 95 scenarios, and the trajectory forecasting dataset, which contains about 80,000 infrastructure-view scenarios, 80,000 vehicle-view scenarios, and 50,000 cooperative-view scenarios captured from 28 intersections' areas, covering 672 hours of data.

News

  • [2023.05] 🔥 V2X-Seq dataset is availale here. It can be unlimitedly downloaded within mainland China.
  • [2023.03] Our new dataset "V2X-Seq: A Large-Scale Sequential Dataset for Vehicle-Infrastructure Cooperative Perception and Forecasting" has been accepted by CVPR2023. Congratulations! We will release the dataset sooner.

Dataset Download

V2X-Seq is one of our DAIR-V2X dataset series. Download links are as follows:

Getting Started

Please refer to getting_started.md for Installation, Evaluation, Benchmark and Training etc for VIC3D.

Benchmark

Coming soon.

TODO List

  • Dataset release
  • Dataset API
  • Evaluation code
  • Training&Evaluation code for VID3D Tracking Benchmark
  • Training&Evaluation code for VID Forecasting

Citation

Please consider citing our paper if the project helps your research with the following BibTex:

@inproceedings{v2x-seq,
  title={V2X-Seq: A large-scale sequential dataset for vehicle-infrastructure cooperative perception and forecasting},
  author={Yu, Haibao and Yang, Wenxian and Ruan, Hongzhi and Yang, Zhenwei and Tang, Yingjuan and Gao, Xu and Hao, Xin and Shi, Yifeng and Pan, Yifeng and Sun, Ning and Song, Juan and Yuan, Jirui and Luo, Ping and Nie, Zaiqing},
  booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
  year={2023},
}
@inproceedings{dair-v2x,
  title={Dair-v2x: A large-scale dataset for vehicle-infrastructure cooperative 3d object detection},
  author={Yu, Haibao and Luo, Yizhen and Shu, Mao and Huo, Yiyi and Yang, Zebang and Shi, Yifeng and Guo, Zhenglong and Li, Hanyu and Hu, Xing and Yuan, Jirui and Nie, Zaiqing},
  booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
  pages={21361--21370},
  year={2022}
}

Contaction

If any questions and suggenstations, please email to dair@air.tsinghua.edu.cn.

Related Resources

Awesome

About