jiangxuehan / RandPerson

Surpassing Real-World Source Training Data: Random 3D Characters for Generalizable Person Re-Identification

Home Page:https://arxiv.org/abs/2006.12774

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

RandPerson

This project contains the ReandPerson dataset described in our paper "Surpassing Real-World Source Training Data: Random 3D Characters for Generalizable Person Re-Identification".

Fig. 1. Sample images from the proposed RandPerson dataset.

Table of Contents

Dataset Description

The RandPerson dataset is generated by MakeHuman and Unity3D. This is the first synthetic person re-identification dataset with a set-up similar to real video surveillance systems, i.e. with a camera network and multiple people moving at the same time. The dataset contains 1,801,816 synthesized person images of 8,000 identities. Images in this dataset generally contain different viewpoints, poses, illuminations, backgrounds, occlusions, and resolutions, as shown in Fig. 1.

Download Links

Due to the large amount of data, currently only the image subset used in the experiments of our paper is provided in Google Drive, including 132,145 images. All data is provided in Baidu Yun Drive, including videos and images (The video is uploading). Images include images and key points position. They can be downloaded from the following links.

File Structure

randperson
├── images
│   ├── scene00
│   │      ├── camera0_1.tar.gz  # This file contains valid pictures in camera0_1.mp4
│   │      │      ├── 000000_s00_c00_f000264.jpg
│   │      │      ├── 000000_s00_c00_f001032.jpg
│   │      │      ├── 000000_s00_c01_f001632.jpg
│   │      ├── camera0_1_point.txt # This file contains key points position of the images in camera0_1.tar.gz. Data format: image name, the upper left corner of the video x, the upper left corner of the video y, the lower right corner of the video x, the lower right corner of the video y, the distance of head point in the image from the upper left corner x (The following distances are from the upper left corner), the distance y of head point in the image, left shoulder distance x, left shoulder distance y, right shoulder x, right shoulder y, left hand x, left hand y, right hand x, right hand y, left foot x, left foot y, right foot x, right foot y
│   │      ├── delete_camera0_1.tar.gz # Because the key points may also be recorded when the model is destroyed, we move the last picture of each id after cutting the picture to the file
│   │      ├── camera0_2.tar.gz
│   │      ├── camera0_2_point.txt
│   │      ├── delete_camera0_2.tar.gz
│   │      ├── camera1_1.tar.gz
│   │      ├── camera1_1_point.txt
│   │      ├── delete_camera1_1.tar.gz
│   │      ├── camera1_2.tar.gz
│   │      ├── camera1_2_point.txt
│   │      ├── delete_camera1_2.tar.gz
│   ├── scene01
│   ```
│   ├── scene10
│   ├── subset
├── videos
│   uploading...

The filenames are encoded as follows. Take "000000_s00_c00_f000264.jpg" as an example,

  • 000000 is the id of the person
  • s00 is the id of the scene
  • c00 is the id of the camera
  • f000264 is the number of frames

Characters and Scenes

We design a method to generate a large number of random UV texture maps and use them to create different 3D clothing models. Then, an automatic code is developed to randomly generate various different 3D characters with diverse clothes, races and attributes. Fig. 2 shows examples of our generated clothing and characters.

          

    Fig. 2. Examples of generated clothes and characters.        Fig. 3. Customized Unity3D scenarios used in this work.

Next, we simulate a number of different virtual environments (including 11 scenes, eight outdoor and three indoor, as shown in Fig. 3) using Unity3D, with customized camera networks similar to real surveillance systems. Finally, we import multiple 3D characters at the same time, with various movements and interactions along different paths through the camera networks. The simulation is illustrated in Fig. 4.

Fig. 4. Example configuration of camera networks and character movements.

Experimental Results

By training person re-identification models on these synthesized person images, we demonstrate, for the first time, that models trained on virtual data can generalize well to unseen target images, surpassing the models trained on various real-world datasets, including CUHK03, Market-1501, and partially DukeMTMC-reID and MSMT17. The experimental results are shown in the following tables.

          

          

Contacts

Yanan Wang
Inception Institute of Artificial Intelligence (IIAI)
yanan.wang@inceptioniai.org

Citation

@article{RandPerson,
author = {Yanan Wang and Shengcai Liao and Ling Shao},
title = {Surpassing Real-World Source Training Data: Random {3D} Characters for Generalizable Person Re-Identification},
journal = {arXiv:2006.12774},
year = {2020}
}

About

Surpassing Real-World Source Training Data: Random 3D Characters for Generalizable Person Re-Identification

https://arxiv.org/abs/2006.12774

License:Apache License 2.0


Languages

Language:Python 100.0%