subburajs / san

The official PyTorch implementation of "Context Matters: Self-Attention for sign Language Recognition"

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

Sign Attention Network

This repository provides a pytorch-based implementation of Context Matters: Self-Attention for Sign Language Recognition. Please Note that in the paper we considered using only the Sign Language Recognition part.

Getting Started

These instructions will get you a copy of the project up and running on your local machine for development and testing purposes.

Updates

  • I am well aware of the code errors. Please use this repository as a reference code as there are things that you may need to change to make it work. Thank you for your understanding.
  • Paper published in ICPR 2020.
  • Paper Arxiv link: https://arxiv.org/abs/2101.04632

Prerequisites

Of course, you need to have python, here we are using python 3.6. So you need to install python3.

sudo apt-get update
sudo apt-get install python3.6

Install Pytorch a very cool machine learning library and the CUDA environment.

https://pytorch.org/

Install opencv2.

sudo apt-get install python-opencv

Other dependencies (numpy, etc..).

pip install numpy

Architecture

Evaluation

To evaluate the SAN model for SLR (Sign Language Recognition)

python evalaute_slr.py

To evaluate the SAN model for SLT (Sign Language Translation)

python evalaute_slt.py

Text simplification

After generating the prediction/translation output texts for the whole text, you can use the above script to remove the unwanted tokens like stop words (This will improve recognition performance).

./simplify.sh <path of the generated texts>

Training

(Optional)You can use dataset_stats to collect stats (std / mean) to better normalize your images.

python dataset_stats.py

To train the SAN model for SLR (Sign Language Recognition)

python train_slr.py

To train the SAN model for SLT (Sign Language Translation)

python train_slt.py

Built With

  • Pytorch - ML library
  • Opencv - Open Source Computer Vision Library

Results

Quantitative Analysis

Qualitative Analysis

Datasets

RWTH-PHOENIX-Weather 2014: Continuous Sign Language Recognition Dataset

https://www-i6.informatik.rwth-aachen.de/~koller/RWTH-PHOENIX/

RWTH-PHOENIX-Weather 2014 T: Parallel Corpus of Sign Language Video, Gloss and Translation

https://www-i6.informatik.rwth-aachen.de/~koller/RWTH-PHOENIX-2014-T/

Contributing

You are free to use this project or contribute that would be cool. Please contact me if you face any problems running the code or if you require any clarification.

License

This project is licensed under the MIT License - see the LICENSE file for details

Authors

Acknowledgments

About

The official PyTorch implementation of "Context Matters: Self-Attention for sign Language Recognition"

License:MIT License


Languages

Language:Python 98.5%Language:Shell 1.5%