SignDiff / Processed-Data

Preprocessed data of SignDiff: Learning Diffusion Models for American Sign Language Production

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Processed-Data

This repository stores the preprocessed data for paper:
SignDiff: Learning Diffusion Models for American Sign Language Production

Note: This work has not been accepted yet, and I am currently very busy and do not have time to submit my paper at a recent academic conference. Maybe it will be accepted at some conference in the fall and that might be a good time to open up the code.

How2Sign for ASLP

After preprocessing How2Sign dataset, the condensed data set obtained is as follows:

It can be used in the training of ASL production models.
Note: Because I later processed more data, the link above is four times the size of the one in the paper and is the result of the full How2Sign processing.

Phoenix-14T for GSLP

After preprocessing Phoenix-14T dataset, the condensed data set obtained is as follows:

It can be used in the training of GSL production models.

How2Sign for SignDiff

After preprocessing How2Sign dataset, the condensed data set obtained is as follows:

It can be used for the diffusion model training of pose2video in sign language. (Based on ControlNet)

How2Sign for Vid2Vid

After preprocessing How2Sign dataset, the condensed data set obtained is as follows:

It can be used for the diffusion model training of pose2video in sign language. (Based on Vid2Vid)

Tool for Data

Our pre-processing tools: the data cleansing tool and the three-step 2Dto3D tool.

Stay tuned. The data above should be sufficient for the time being.

Related Work

About

Preprocessed data of SignDiff: Learning Diffusion Models for American Sign Language Production


Languages

Language:Python 100.0%